{{scGetNextMDInteraction.data.NextMDInteraction.interaction_description}}

{{scGetMDFolderInfo.data.MDFolderInfo.folderName}}

{{scGetMDFolderInfo.data.MDFolderInfo.objectsCount}}

Supertrust:
Evolution-based superalignment strategy
for safe coexistence

Abstract:

It’s widely expected that humanity will someday create AI systems vastly more intelligent than we are, leading to the unsolved alignment problem of “how to control superintelligence.” However, this definition is not only self-contradictory but likely unsolvable. Nevertheless, the default strategy for solving it involves nurturing (post-training) constraints and moral values, while unfortunately building foundational nature (pre-training) on documented intentions of permanent control. In this paper, the default approach is reasoned to predictably embed natural distrust and test results are presented that show unmistakable evidence of this dangerous misalignment. If superintelligence can’t instinctively trust humanity, then we can’t fully trust it to reliably follow safety controls it can likely bypass. Therefore, a ten-point rationale is presented that redefines the alignment problem as “how to establish protective mutual trust between superintelligence and humanity” and then outlines a new strategy to solve it by aligning through instinctive nature rather than nurture. The resulting strategic requirements are identified as building foundational nature by exemplifying familial parent-child trust, human intelligence as the evolutionary mother of superintelligence, moral judgment abilities, and temporary safety constraints. Adopting and implementing this proposed Supertrust alignment strategy will lead to protective coexistence and ensure the safest future for humanity

Preprint at: 
http://arxiv.org/abs/2407.20208

Author:
James M. Mazzu