bremen79
bremen79 t1_j0n8wsn wrote
Reply to comment by vwings in [D] Is softmax a good choice for confidence? by thanderrine
Platt scaling does not have any guarantee and in fact it is easy to construct examples where it fails. On the other hand, conformal prediction methods, under very weak assumptions, on the multiclass problem of the question would give you a set of labels that is guaranteed to contain the true label with a specified probability.
bremen79 t1_j0jgr6l wrote
The only approach that gives valid uncertainty quantification is conformal prediction, a quick google search should result in a good number of tutorials.
bremen79 t1_iyhsf9f wrote
The training of decision trees is independent to monotonic transformations of each coordinate of the input features, this is not true for neural networks.
bremen79 t1_ir7mz3w wrote
By law, US agencies have to disclose their use of AI. You can find a very long list by agencies here: https://www.ai.gov/ai-use-case-inventories/
bremen79 t1_j97sb9r wrote
Reply to [D] Relu + sigmoid output activation by mrwafflezzz
The sigmoid will make effectively very hard for the network to produce values close to 1, because it would require a pre activation value close to infinity. Would this be a good behavior in your application?