Viewing a single comment thread. View all comments

rushmc1 t1_j6ov99k wrote

Yeah...that's not how it works.

0

turnip_burrito t1_j6ovrva wrote

Is it? There is a new Google robot (last couple months) that uses LLMs to help build its instructions for how to complete tasks. The sequence generated by the LLM becomes the actions it should take. The language sequence generation determines behavior.

There was also someone on Twitter (last week) who linked chatGPT to external tools and the Internet. This allowed it to solve a problem interactively, using the LLM as the central planner and decision maker. Again here, the language sequence generation determines behavior.

Aside fron these, alignment is the problem of controlling behavior, and behavior is a sequence. The rules and tricks discovered for controlling language sequences maybe can help us understand how to control the larger behavior sequence.

Mostly just thinking aloud. Maybe I'm just dumb, since everyone here in the comments seems to have the opposite opinion of mine, but what do we make of the two above LLM use cases where LLMs determine the behavior?

1