Motivational Defeaters of Self-Modifying AGIs
The future of artificial general intelligence (AGI) in general and the risks and promises it may hold in particular are the subject of many recent papers and books by many eminent thinkers. In this paper I discuss a widely held argument in this field of enquiry, i.e. the Intelligence Explosion Argument (IEA), and argue that motivational defeaters render the IEA unsound. I discuss the argument from the prism of self-modifying artificial intelligent systems and demonstrate how certain considerations might cause them, under certain circumstances, to be disinclined to perform self-modifications. The argument I promote in this paper is simple: self-improvement is a risky process. Our AGI is intelligent enough to understand this. Therefore, under normal circumstances, it will be disinclined to self-improve. By normal circumstances I mean circumstances wherein the successful completion of its ultimate goals is not at risk. Since the IEA requires a certain recursive self-improvement on the part of the agent, disinclination to do so will render the IEA unsound.
No Supplementary Data
No Article Media
Document Type: Research Article
Affiliations: Email: [email protected]
Publication date: January 1, 2017