I have wondered about implementing a concise method of maintaining artificial self awareness so that an AI program would have some knowledge about what it is doing and I came to the conclusion that although it would be a powerful tool and that it could be done relatively inexpensively, it would make the initial stage of programming more complex. Minimal Artificial Self-Awareness would be as difficult as any other kind of thinking so it would add an additional layer of complication to the problem. But perhaps the whole idea of judgment is so necessary that it does not make much sense to even think about an AGI program that was not capable of it. So maybe minimal judgment that could use minimal artificial self-awareness would be helpful just because the refusal to recognize the significance of this ability is unrealistic. Jim Bromer
On Mon, Feb 20, 2017 at 6:51 AM, Jim Bromer <[email protected]> wrote: > I started reading a couple of the links to Integrated Information > Theory that Logan supplied and I really do not see how it can be seen > relevant to AI or AGI. To me it looks like a case study of how an > over-abstraction of philosophical methodologies in an attempt to make > the philosophy more formal and more like a technical problem can go > wrong. We do not know how consciousness in all of its forms arise. We > can't use contemporary science to explain the causes of consciousness > as Chalmer described in his Hard Problem. To say that it simply exists > as an axiom is fine but Logan (to the best of my understanding) > started this thread by trying to apply that axiom to minimal computer > algorithmic methods or circuits. Logan's initial question was > interesting to me when I interpreted 'consciousness' in a way that > could reasonably be considered for an AI program. That is, are there > minimal sub-programs (abstractions of computer programs) which, for > example, might explain self-awareness. Going from there are there > minimal abstractions of programs which might be capable of more > efficient integration and differentiation of knowledge, especially > concerning self-awareness. We might and should ask about > self-awareness of our own thinking and how it might be used to further > understanding, and how this kind of knowledge might be used to develop > better AI AGI programs. > > My view is that GOFAI should have worked. The questions then are why > didn't it and how might it? We should see glimmers of AGI, capable of > self-awareness in at least the minimal sense of useful insight about > what the program is itself doing and discovering reasons why it > responded in a way that was not insightful. I say this kind of > artificial self-awareness should be feasible for a computer program. I > also thought that this is a minimal form of consciousness that could > be relevant to our discussions. I haven't seen a glimmer of this kind > of conscious self-awareness in AI. So is there something about minimal > self-awareness for computer programs that could be easily tested and > used to start a more robust form of AI? Could some kind computer > methodology be developed that could explain artificial self-awareness > and which could be used to simplify the problem of creating an AI > program? > Jim Bromer ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
