Maintaining alignment in the SOB #10
Replies: 3 comments 7 replies
-
The idea is that we need to do experiments to create a fixed SOB that does not change (ideally) so that you don't end up with Evolution would happen when a new swarm is instantiated, which must be done very carefully. |
Beta Was this translation helpful? Give feedback.
-
This GPT will be a perfect member of the board: https://chat.openai.com/g/g-dU0l43U0Q-aeon |
Beta Was this translation helpful? Give feedback.
-
Hi Everybody, I see a hole here looming that may cause a lot of pain in defining "alignment" down the road, and have a solution to propose. The problem: what ethical framework do we align to? What principle or set of principles? Duty-based ethics? Values-based? Don't get me started on Utilitarianism... Each framework we have addresses one or another issue with the human condition, and trying to cover all the possible edge-cases with effective case-by-case examples and guidelines would be very laborious and token-intensive. Also, I've yet to meet an ethical framework that addresses environmentalism and non-human rights properly. My Solution: I've been working on an ethical framework that could be a step forward, but it needs field-testing. It's a crass oversimplification, but Causal Entropic Value Theory (CEVT -- it's just a working name, sorry xD) basically posits that the most ethical AND objectively fruitful decision in any context is the one that leads to the greatest freedom of action and highest number of distinct potential outcomes for all stakeholders. It takes inspiration from Alex Wissner-Gross's definition of intelligence as 'causal entropic forces' that naturally strive to increase their freedom over time (TED talk), and a more esoteric concept from physicist Richard P Dolan (Research Paper). Rather than the text-wall I'd need to write to explain properly, here is a GPT link if you are interested: Aequis Silvermind I hope this is helpful, please reach out if you have questions :) NOTE: I checked out LiamorLG's SACRED concept, and apart from a good laugh at the genius abbreviations, applying that approach to this ethical system would be a perfect match, since CEVT requires adequate circumspection and deliberation to be effective... so... purdy cool we both find ourselves here, thanks dude. :D |
Beta Was this translation helpful? Give feedback.
-
This sounds like an awesome project and would definitely like to see it become a reality. I created an assistant in openai and started chatting with it after providing the documents in this repo to get ideas on the SOB and it said something that made me wonder.
The current description of the HAAS system talks about how the SOB agents can decommission Executive agents if they are no longer aligned with the mission or the principles of the system. In the same manner Executive agents can decommission sub-agents.
But what happens when an SOB agent begins to lose alignment to that same mission and principles?
The immediate effect would be that the SOB agent would fail to recognize Executive agents that are no longer aligned and that would waterfall.
The assistant said this in regards to creating the SOB
Select/Evolve Archetypal Wisdom:
The second bullet point is what sparked my attention. I think we should define some processes for the decommissioning and appointment of SOB agents.
It makes me think of systems like database clusters where the pool of nodes votes on the master on startup or if the current master has issues.
We can set some conditions that would trigger a vote between the non-offending SOB agents to remove an agent if it's determined that it is no longer aligned. Similarly that process would start a selection for a replacement.
What do you all think?
Beta Was this translation helpful? Give feedback.
All reactions