Submitted by ken81987 t3_10k0fx0 in singularity
turnip_burrito t1_j5o7uss wrote
Reply to comment by cloudrunner69 in how will agi play out? by ken81987
The idea is a result of the "orthogonality thesis": the idea that goals and intelligence are two separate aspects of a system. Basically, a goal would be set and the intelligence is just a way to achieve the goal.
This kind of behavior is seen in reinforcement learning systems where humans specify a cost function, which the AI minimizes (equivalently maximizing reward). The AI will act to fulfill its goal (maximize reward) but do stupid stuff the researchers never wanted it to do, like spinning in tiny circles around the finish line of a racetrack to rack up points, for example. It's the same kind of loophole logic that comes from stories of lawyers, genies, and such that the AI agent uses to maximize reward.
It's entirely possible this method of training an agent (maximize this one loss function) is super flawed and a way better solution is yet to be created.
Viewing a single comment thread. View all comments