avialex

avialex t1_iu6ajry wrote

I use fullgrad religiously, although I've removed the multiplication by the original image so that I'm just seeing the model gradients. I don't really use it to debug, it's more useful as a post-facto indication of what the important features in the data were. Every once in a while I'll see a model is overly focused on corners or something obviously wrong, and that can be an indication of too much instability, but aside from that it's more of an explanatory tool than a debugging tool.

1

avialex OP t1_irnpnry wrote

Quantum NN's are quantum algorithms, are they not? Are you thinking of hybrid nets where only a few neurons are quantum?

edit: ok I see, you're saying GD is the problem, we need a QC algorithm to train QNN's. I would definitely agree, but as it stands I don't think there is one?

−1

avialex OP t1_irnnm6a wrote

They certainly are looking, but at the same time gradient calculation is fundamental to how quantum neural networks are implemented right now, and QNN's are a relatively active area of study. I don't think we can dismiss the work in the field as it stands, because it's all built on the foundation of gradient descent. Afaik no one has yet found a better way to train a QNN, even on quantum data. I could be wrong.

4

avialex t1_ir41bgt wrote

"The model is queried to generate modifications of an initial source code snippet. In our experiments, this is a network with a single hidden layer of 16 neurons. The possible modifications include adding convolutional layers, changing the size of convolutional or hidden layers, and increasing the number of hidden layers."

Lmao...

How about those smooth curve lines on graphs with fewer than 10 sample points? That inspires confidence.

24