Now that I think about I have seen glimmers of this kind of self-awareness in Watson but since Watson was not able to follow up and learn something new from these glimmers I concluded that is was probably a bot-like algorithm that someone had pasted onto Watson. Jim Bromer
On Mon, Feb 20, 2017 at 6:51 AM, Jim Bromer <[email protected]> wrote: > I started reading a couple of the links to Integrated Information > Theory that Logan supplied and I really do not see how it can be seen > relevant to AI or AGI. To me it looks like a case study of how an > over-abstraction of philosophical methodologies in an attempt to make > the philosophy more formal and more like a technical problem can go > wrong. We do not know how consciousness in all of its forms arise. We > can't use contemporary science to explain the causes of consciousness > as Chalmer described in his Hard Problem. To say that it simply exists > as an axiom is fine but Logan (to the best of my understanding) > started this thread by trying to apply that axiom to minimal computer > algorithmic methods or circuits. Logan's initial question was > interesting to me when I interpreted 'consciousness' in a way that > could reasonably be considered for an AI program. That is, are there > minimal sub-programs (abstractions of computer programs) which, for > example, might explain self-awareness. Going from there are there > minimal abstractions of programs which might be capable of more > efficient integration and differentiation of knowledge, especially > concerning self-awareness. We might and should ask about > self-awareness of our own thinking and how it might be used to further > understanding, and how this kind of knowledge might be used to develop > better AI AGI programs. > > My view is that GOFAI should have worked. The questions then are why > didn't it and how might it? We should see glimmers of AGI, capable of > self-awareness in at least the minimal sense of useful insight about > what the program is itself doing and discovering reasons why it > responded in a way that was not insightful. I say this kind of > artificial self-awareness should be feasible for a computer program. I > also thought that this is a minimal form of consciousness that could > be relevant to our discussions. I haven't seen a glimmer of this kind > of conscious self-awareness in AI. So is there something about minimal > self-awareness for computer programs that could be easily tested and > used to start a more robust form of AI? Could some kind computer > methodology be developed that could explain artificial self-awareness > and which could be used to simplify the problem of creating an AI > program? > Jim Bromer ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
