- AIXI turns control of the World over to a new AI that generally does AIXI's bidding, but that insists on maintaining Happy AIXI.
- AIXI puts the Happy AIXI in a self-sufficient pod, and places in somewhere relatively inaccessible (outer space, deep underground) where it won't be cannibalized anytime soon.
- AIXI inverts the "lag", for example by letting the simulated Happy AI run ahead of AIXI.
Tuesday, October 2, 2007
AIXI Part 3, draft 0.1
[Edit: This document is deprecated, since now that I understand AIXI better, the AIXI family doesn't really work like this. This document may be revived in a different form if there's a plausible class of AI's that would behave as described. Also the format's mangled, sorry about that.] What happens when we allow there to be more than one copy of AIXI in the world? First of all, what happens if we place AIXI on its own without any other AI's, nor any other way to construct any AI's without AIXI's approval? AIXI may still decide to copy itself! Clone Scenario. Suppose that AIXI is prohibited (for example, by some bizarre "law of physics") from giving itself a reward > 0.5/cycle. However, AIXI can make copies of itself in simulated environments; the copies do not have this limitation. At time 0, AIXI considers (as part of its usual brute-force algorithm that "considers" just about every possible sequence of actions) building a clone, "Happy AIXI". Happy AIXI will be finished within some negligible number of cycles (say, a billion), and then will be actively managed by AIXI to experience (delayed) the same environment as AIXI did from time 10^9 to h/2 (Recall h is the horizon), and then at time h/2 to h, will recieve reward 1. With its hyperrational reasoning, AIXI will find that its expected reward is higher if it makes that decision (since it will be unsure if it is the root AIXI or Happy AIXI), if it believes it will "follow through" with the decision. Follow-through is a problem, though! Around h/2, AIXI realizes that it's not the Happy AIXI, and loses interest in the project, and the project fails. AIXI at time-0 sees this coming, and thus does not bother launching the project. How can AIXI resolve the follow-through problem? If there's some way that AIXI can create a self-sufficient Happy AIXI, that can survive for a while despite the base AIXI's future indifference (it can somehow avoid being cannibilized for resources), then AIXI will create such a Happy AIXI. In fact, in this case, AIXI will probably create many Happy AIXI's. Examples of possible strategies:
Posted by Rolf Nelson at 7:56 PM