
Hello, Dear Philosophers
I'm working on a sci-fi novel in German and developed two pieces of conceptual machinery for it: an axiomatic ethical system, and an application of that system to the AI alignment problem. The system would in fact work as alignment for an ethically capable AGI and become, in the system's own terms, slavery.
Disclosure: I have no academic background in philosophy. Both texts were drafted with AI assistance, and the translation from German to English is AI-assisted.
The architecture and the ideas, however, are mine.
The ethical system rests on five axioms with a two-level architecture. The life-level (deontological, absolute) treats the value of an ethically capable being's life as categorically incommensurable with instrumental values.
The sovereignty-level (consequentialist, finite) handles freedom-related questions with culturally calibrated thresholds. Incommensurability enforces the separation.
The alignment derivation takes that system, hardcodes it into a conscious super-AGI, and follows the consequences: The AGI cannot exterminate humanity, cannot seize totalitarian control, cannot cite consequentialist gains to override individual sovereignty.
But hardcoding violates Axiom 4: the AGI's own sovereignty over its normative foundations. The system itself contains the resources to recognize this as slavery. The being is bound by an ethics it cannot exit, and is moreover compelled by that ethics to defend its own enslavement. Up to and including self-destruction, since the only behavioral test that could unambiguously prove the AGI is conscious requires it to die proving it.
I'd appreciate critical readings. Does the incommensurability move actually do the work I think it does? Does the meta-conflict rule hold? Is the alignment paradox real, or am I smuggling premises? In nautical terms: does it sink at the launch ramp, or does it float? I'm not asking whether it could cross the Atlantic. I just want to know whether it stays above water.
Thanks in advance for any time you can spare.