Viewing a single comment thread. View all comments

newperson77777777 t1_j1qv2cz wrote

You may want to check for bugs as well with your normalization process. Additionally, generally, if you're doing standard normalization for images, you do it over the training set rather than per image. So you get the mean and standard deviation over your training set and calculate the normalized images. However, you seem to suggest that there are some differences in the images depending on the year and other qualities. Thus, you may want to standard normalize by year or something else.

I'm assuming this is a segmentation problem? How do you know the model is memorizing the input range and not learning the spatial patterns? If there's a memorization issue, the model may be overfitting - thus, you may want to consider using data augmentation or more data or try simpler models.

6

skn133229 OP t1_j1r2mfy wrote

Yes this is a segmentation problem. I have a preprocessing augmentation function that randomly rotate, flip and translate the images. My network also starts with a gaussiannoise layer which adds some random noise to the input before entering the unet network. I thought it was memorizing input ranges because of the poor performance when images are normalized individually. When I peak at the normalized input images I see the salient patterns I was hoping that the unet would latch onto but training fails. I can rule out bug in the normalization process because I can visualize the normalized input images and they look fine.

2

newperson77777777 t1_j1r9oni wrote

So based on my understanding, training fails when you try to normalize per image, right? When you don't normalize per-image, training is fine but you just get suboptimal validation performance - which, in this case, you are saying is because there may be more error during certain years, especially years that are not considered during training. Is that an accurate summary?

If this is the case, I would try to test two things independently (not at the same time): 1. standard normalization over the entire dataset and 2. standard normalization per year. I would also explore the data based on year year and try to see what are the differences to see how you can adjust your modeling. Additionally, I would try do a more thorough error analysis to understand patterns in the errors the model is making.

3

skn133229 OP t1_j1t9qgo wrote

I solved the problem with failure to train (see update in original post). It appears that implementing some random spectral shifts and random noise were necessary for training to take off. I will try your suggestions to see if I get better prediction results year over year.

1

newperson77777777 t1_j1tbh2m wrote

Weird, haven't encountered that issue before. If you do the feature-wise normalization that I discussed before, more likely than not that will result in, for most images, the image channel mean not being 0 as well, if that helps.

1