killver

killver t1_iv0ycz4 wrote

If you trust a random blog, go ahead.

This ruling was for a very specific use case that cannot be generalized, and also only applies to US, even only a specific district. It is also totally unclear how it applies to generative model, which even the blog cited recognizes.

The AI community just loves to trust this as it is the easy and convenient thing to do.

Also see a reply to this post you shared: https://medium.com/@brianjleeofcl/this-piece-should-be-retracted-ca740d9a36fe

5

killver t1_itpw963 wrote

The easiest way that works well in practice is to just concatenate them. You can also normalize them first separately before concatenation. If one dimension is significantly different, you can just concatenate the other one multiple times to weight them similarly, or use a dimensionality reduction beforehand.

Another way is to just calculate two similarities separately and then average them (or weighted average).

You can take a look at this kaggle competition's solutions for inspiration: https://www.kaggle.com/competitions/shopee-product-matching/discussion

3

killver t1_iqo6z39 wrote

Alpha in focal loss has confused me and others before. I do not understand why they built their paper writeup so heavily around it, as it was not really the contribution of the paper.

I would suggest to use a non-alpha variant in your experiments, and only think about alpha as a common way of up/downscaling classes and add it later.

2