mr_birrd

mr_birrd t1_j77rkjd wrote

If a LLM model tells you it would rob a bank it's not that the model would do that could it walk around. It's what a statement that has a high likelihood in the considered language for the specific data looks like. And if it's chat gpt the response is also tailored to suit human preference.

3

mr_birrd t1_j6xtb3u wrote

Edit: Chatgpt uses GPT3. Search the dataset it used.

Google it they have full transparency. If you find a text by yourself there maybe ask if they can remove it. First of all, the data is only used for stachastic gradient descent and the model has no idea about the content it read, it only can model probabilities of words, e.g. it learned to speak but it only speaks such that it mostly outputs what makes sence in a bayesian way.

So the model is already trained and it didn't even read all of the data, those huge models often only read each instance of sample once at maximum, since they learn that "well".

Also in the law text you wrote I understand it that if you opt out in the future, it doesn't make past data processing wrong. The model is already trained, so they don't have to remove anything.

They also mostly have a whole ethics chapter in their papers, maybe you go check it out. Ethics etc is not smth unknows and especially such big companies also have some people working on that in their teams.

1

mr_birrd t1_j6xo5qk wrote

No it doesn't raise ethical concerns. You literally have to agree about usage about your data and at least in Europe should be able to opt out of everything if you want. You should 100% know this, those are the rules of the game. Just cause you don't read the terms of agreements doesn't make it unethical for companies to read your data. Sure if you then use it for insurances that won't help you cause you will become sick w.h.p. that's another thing. But don't act surprised.

1

mr_birrd t1_j6f7h26 wrote

Well like reinforcement learning uses a lot of markov chains, forward/backwards filtering/smoothing etc. Kalman filters are also a sort of Gaussian Process Regression. There is a huge overlap in the classical ML part with signal processing. No specific paper but it's just that ML and especially deep learning often takes already existing ideas from physics or ee and try to apply it on some data, see what happens.

1

mr_birrd t1_j2s38gv wrote

Which will lead to ultra high taxes in europe and is even legal in some(if you work both 100% without the other noticing). I don't see a problem, nearly noone would do that where I live, free time > money (at least knowing that the income from one ML job is probably quite enough).

1

mr_birrd t1_j06s0mg wrote

Well uint 8 goes to 255, so there you take those values. Images come in that format often but the ReLUs and other activations hate it so better take it to a 0-1 range. Btw min max just subtracts the min of the sample and then divide by max. I don't see the problem.

Edit: Also think about why we do BatchNormalization

−1