9 Jul
2023
9 Jul
'23
10:18 p.m.
maybe it bears some similarity to that turing problem of predicting the completion of another code turing’s contrived counterexample places power regarding prediction and control in such a way that the goal cannot succeed. of course this proof also disproves itself in some ways because the code in question must be able to predict the behavior of the halting-prediction code. in a normal, realistic scenario, it’s possible to predict the behavior of a system in most situations, and in many it’s unreasonable to do so, but you can in a lot of them if you work really hard to. one of these difficult situation is when the system is observing your prediction, and attempting to be unpredictable — or otherwise pursuing a conflicting goal.