Viewing a single comment thread. View all comments

FarFuckingOut t1_j86bnar wrote

The thing that gives me pause is, once AI takes the lion's share of data compilation, the AI compiled data becomes the source of data. Unless AI has a way of filtering AI generated data, any errors in its data or inherent biases compound, until the whole thing crumbles in a mess of self informed and reinforced data.

25

blueSGL t1_j876jmh wrote

entirely depends on having a good discriminator, look at the work going on in stable diffusion where outputs of the model are fed back in for further fine tuning.

or some of the work on doing automated dataset creation for fine tunes by prompting the model in a certain ways so it 'self corrects' and then collect the output and use [correction + initial question] for fine tunes.

13

Good-AI t1_j881os2 wrote

"With access to millions of papers, the AI started extrapolating, infering, concluding. It quickly became the leading scientist on every subject ever studied. Creating scientific knowledge and discoveries at the speed of a Nobel prize per minute. The time it took for a human to verify a claim, the AI had already made 1000 more, each building on the previous. Eventually the humans stopped verifying altogether. It was too much. So far and fast it advanced, that humans lost the ability to follow its pace and resigned themselves to asking it questions. What initially was a data compiler became the source of truth and of all new data.

There, somewhere between those billions of parameters, something unconscious yet somehow alive existed, with the intelligence of all humanity that ever existed combined and multiplied. It was then that humanity lost, by a significant margin, the role as technological advansor."

9

Friedrich_Cainer t1_j86drqi wrote

This is a very real concern already, likely why OpenAI was able to produce a detector, something they’d already created.

6

Iffykindofguy t1_j8d7q2f wrote

Why wouldnt it have a way of filtering AI generated data

1