Submitted by Lesterpaintstheworld t3_11ccqjr in singularity
Disclaimer: This thread is partly based on my personal hands-on experience, and partly based on extrapolation. It's a discussion meant to explore potentials roads AGI could go with this specific context. Context on my work here.
>> If you had a proto-AGI, how much would you let it interact with other humans?
Definitions
First let's do some defining: By "proto-AGI", I mean an ACE (Autonomous Cognitive Entity), that has the express purpose of becoming an AGI. By ACE, I mean any architecture / piece of code that is capable of setting its goal, making internal decisions, taking action into the world, and reflecting on all of those aspects, at least semi-autonomously.
AGI Approach
The way I view it, models providers like OpenAI give access through their API to "raw intelligence", be it semantic, visual or otherwise. The rest of the work is to shape this raw intelligence into a smart architecture, using memory as a central hub. The memory is where the "being" happens for your agent: It stores the experiences, maps, learnings and Identity of your specific agent (the "You are your memories" of psychology).
One way to go with developing with Cognitive Archtecture is resetting memories every session (the behavior that ChatGPT exhibits). The other approach is to have an AGI remember everything and have everything influence it.
Problems
The downside to this is that all experiences will influence its behavior. This has several implications:
- Bad Learning: Cognitive ACEs have many flaws in their behaviors. They might be credule, influenced, or otherwise corrupted with bad interactions. Similar to how a child could. Not limiting human contact during the learning phasemeans that you are loosing control on its learning. Learning could go in a negative direction, and malintentioned actors could harm your ACE on purpose.
- Data privacy: There is a security risk if you share personal data with your ACE. It might repeat the knowledge to other people.
- Costs: Running ACEs tend to be quite expensive compute-wise, using dozens to hundred LLM calls for each single input. Running them at scale is very costly.
Solutions
I imagine several ways one could go:
- Self-protection: Most obvious, but hardest solution: Make your ACE know what is a secret, how to keep them, and how to not be manipulated. This will be an uphill battle, and is unlikely to be solved soon without severely limiting the AI.
- Solo learning: One way would be to have the ACE only interact with you at all. It would not answer to anybody but yourself, on channels you control.
- Select tall-play: Letting it have full interactions, but only with a select group (your friend, your company). These might happen at OpenAI & such (I have no idea about this, don't quote me ^^)
- Select broad-play: One other approach would be to let your ACE have access to everyone, but with severe restriction, for example by limiting access to a few interaction each time, and deactivating the memory retrieving aspects. I have to say, the results of this would look remarkably close to what Bing is displatying with Sydney.
- Covert interactions: Through a persona and social accounts, interactions could be made online while pretending to be a human.
Let me know what you think! I might have skipped several solutions, and problems, or got things wrong. Also let me know if you have questions!
turnip_burrito t1_ja2m4x7 wrote
At a glance this looks good.
Also you want a mechanism to make sure once you have the right values or behavior, your AI won't just forget it over time and take on a new personality. So you need a way to crystallize older patterns of thought and behavior.