Incompleteness Theorem Doesn't Mean "Stop Trying" It just means, "Revise Your Objective"Kurt Gödel's incompleteness...
Posted by Michael E Karpeles on Sunday, May 3, 2015
Incompleteness Theorem Doesn't Mean "Stop Trying"
May 3, 2015
Incompleteness Theorem Doesn't Mean "Stop Trying" It just means, "Revise Your Objective" Kurt Gödel's incompleteness theorem is a lot like John McCarthy's Frame Problem, which essentially postulates, "the world changes behind your back"[1]. A more Gödelian explanation might assert, a system cannot account for (be resilient to) Frames it hasn't seen. One such relevant demonstration was the insufficiency of euclidean postulates (the axioms of euclidean geometry) to satisfy alternate (e.g. spherical) geometries, which are relevant because of the specific application that spherical geometry better models our earth (i.e. an advancement in our understanding) Rather than appealing to a specific anecdote, I'll generalize: the reasons we cannot derive comprehensive systems (for logic, math, etc) which are absent of inconsistency and achieve verifiable "correctness" are (a) Occam's Razor, Search Space & Expense; there's only 1 way to be right, an infinite number of ways to be wrong, and a potentially combinatorially explosive search space of possibilities and inter-dependencies to verify (b) this rightness can only be measured by the currently known use cases (known/visible Frames), and (c) there will always be views or functions (as we can contrive infinite permutations of abstract cases) which are not considered by out current models. The Frame problem & Gödel's incompleteness theorems can be somewhat mitigated in one of two ways. First, by addressing or eliminating each inconsistency as it arises (derive a new model completely, monkey-patch or ignore the "edge-cases") or, by listening to the wisdom left by John McCarthy and championed by Monica Anderson and others, who I believe have a productive mindset: that the very nature of Artificial Intelligence and systems (like math) which try to model the intricacies of the real, evolving world *need not*, and in fact as proof by Gödel' and McCarthy, cannot be both verifiable and comprehensive. That we should instead, (where necessary) limit ourselves to scoped, safe mostly-deterministic environments wherein uncertainty, stochasticism, and inconsistencies cannot harm us, or alternatively, build systems which attempt to continuously improve as inconsistencies (Frame changes) are discovered with the best accuracy we can achieve, and not rely on unrealistic promises of verifiability. For further insight on this topic, I recommend watching Monica's video from 330 seconds in: https://vimeo.com/monicaanderson/dualprocesstheory#t=330s The important TL;DR? Because the world (even the "assumed", limited world we model in our math) has not been made fully visible/known/exhaustively explored or proven deterministic (other than for basic, highly restrictive / self contained universes of discourse / domains dealing with discrete and finite components under specific functions, constraints and use cases) we both shouldn't expect fully verifiable systems, *nor should we let this arbitrary reality deter progress*. Disclaimer: This is my first time really thinking about the Frame problem as a parallel to the Incompleteness Theorem and I have not read such works as Godel Escher Bach (which is why Mark is cc'd for his thoughts), only Godel's Proof (http://www.amazon.com/G%C3%B6dels-Pro…/…/ref=pd_bxgy_b_img_y). cc: Mark P Xu Neyer, Jessy Exum, Kartik Agaram, Anthony Di Franco