I have seen that Google's AutoML will use some pre-training image augmentation to increase the robustness of the model. (Adjacent discussion)
I have searched the documentation and forums for a way to limit these techniques. For instance it applies flips to the objects. However, in some cases flips hurt the predictions. For instance recognizing numbers in an image. For most fonts, 2's and 5's are different enough to have different features, even when flipped. However, a 7-segment display will have the same representation for 2's and 5's when they are flipped. 7-segment display example
I have labeled hundreds of images with many digits in each image. The model continues to confuse the 2's and 5's for the 7-segment displays. It has some success but not an acceptable amount.
Does anyone know if limiting the image augmentation with AutoML is possible?