Smart person, nanograte, one of the smartest nanogrates I know... I like this:
"Herein lies the deabstraction logic, in the ability to reliably and consistently make sense of how things systemize at any class of abstraction (conceptual, and/or logical and /or physical), then to adaptively progress (read evolutionary step) to another level of abstraction thereafter (repeat: conceptual, and/or logical, and /or physical) on the basis of such sense-making, but not on that basis alone. " My view of AGI is holistic, and I like thinking in terms of abstracting from the whole -- as opposed to something like manipulating discrete symbols and propositions etc. But "deabstraction" conjures up images of some kind of calculus operation on the whole idea. On 4/12/17, Nanograte Knowledge Technologies <[email protected]> wrote: > "That is the abstraction problem of AGI. For example, is health a more > significant domain than finance? Is public service better for the AGI than > bettering the skills of the AGI?" > > I see no abstraction problem with AGI. The examples you posed as problems > are fairly easily resolvable via existential logic. "Domain Provable" > probabilistic choices flow from existential logic. That is where a sense of > correctness is born in the mind of humans, and it could be so in > computerized machines also. And once sense is established, consciousness > becomes possible. However, it does return me to the obvious need for an > adequate, <deabstraction> methodology. > > > "The issue is, where will AGI get the assumptions? And, how rigorous will > the process be for accepting a new assumption?" > > The relative terms 'right' and 'wrong', 'good' and 'bad' etc. carry their > own poison. I prefer to use the term 'correct', to relate a decision to a > scenario option. Indeed, 'correct' also denotes a judgment, but it more > strongly denotes the testable outcome of an assumption, relative to a > knowledge base. > > AGI would get its assumptions from learning, per contextual schema, what a > scale of correctness would result in. Instead of just the two poles of > 'correct' and 'not correct', many other points of correctness could be > defined and placed on such a scale to introduce decision granularity, and so > increase the overall probability of an assumption becoming testable relative > to reality. > > How rigorous will the process be for accepting a new assumption? Not > rigorous at all. The "most true, or most correct" result would always inform > the validity and reliability of any assumption. The strongest genes would > survive. > For example: Start - logic, then assumption, then else chain > reaction. The value on the "correctness" scale would provide the > loop-until value <x>. Exit. End. > > ________________________________ > From: Nanograte Knowledge Technologies <[email protected]> > Sent: 12 April 2017 09:29 PM > To: [email protected] > Subject: Re: [agi] I Still Do Not Believe That Probability Is a Good Basis > for AGI > > > "Okay, but this begs the question of how you define AGI. Domain knowledge > is the distinguishing point of what might be called regular AI. It is the > General part of AGI that doesn't allow a domain intense approach." > > I do not have my own definition of AGI. Any accepted definition is fine by > me, but I understand AGI to mean that a computerized machine would be able > to exhibit human functionality via human-like brain functionality, as > sentient intelligence. In the main, domain knowledge pertains to knowledge > about any domain. Knowledge to me is not AI, but it could be argued to be > so. To me, AI is reasoning towards knowledge. > > On the contrary, I would contend that it is exactly the General part of AGI, > which most allows for a domain intense approach. If we replaced the broader > term 'domain', with a more specialized term, 'schema', and expanded it to > specifically mean 'contextual schema', would your argument still hold > equally strongly? > > > ________________________________ > From: Stanley Nilsen <[email protected]> > Sent: 12 April 2017 05:16 PM > To: AGI > Subject: Re: [agi] I Still Do Not Believe That Probability Is a Good Basis > for AGI > > On 04/11/2017 10:00 PM, Nanograte Knowledge Technologies wrote: > > The moment relationships of any functional value (associations), and any > framework of hierarchy (systems) can be established and tested against all > known (domain) knowledge, and even changed if the rules driving such a > hierarchy should change (adapted), it may be regarded as a concrete version > of a probabilistic framework. > > > Okay, but this begs the question of how you define AGI. Domain knowledge is > the distinguishing point of what might be called regular AI. It is the > General part of AGI that doesn't allow a domain intense approach. > > Is it accepted that the "general" indicates that we are looking across > domains into the realm of all domains? And, we have to choose between > actions coming from multiple domains. One might call this "meta-domain" > knowledge. Such knowledge, I believe, would require abstraction. That is > the abstraction problem of AGI. For example, is health a more significant > domain than finance? Is public service better for the AGI than bettering > the skills of the AGI? Choices, choices, choices... > > > To contend: Probability may not be a "good" basis for AGI, similarly as love > may not be a good basis for marriage, but what might just be a "good" basis > is a reliable engine (reasoning and unreasoning computational framework) for > managing relativity with. This is where philosophy started from, unraveling > a reasoning ontology. > > > I don't think probability is a problem. A piece of knowledge may increase > the chance that we see the situation accurately, and accuracy will help us > be more specific about our response. That said, it is the way we put > assumptions together that will determine our final action. > > Probability has been used in that we think our assumptions are "probably" > right. It is the qualifying of our assumptions that distinguishes the > quality of our actions. Adopt sloppy assumptions and your results will > probably not always be appropriate or best - not super intelligent. > > An "advanced" system will have some mechanism for adopting assumptions (most > currently rely on the judgment of the programmer.) It is this process of > evaluating assumptions that we tend to get abstract. Since we are calling > these "heuristics" assumptions, there is an implication that we can't prove > this premise that we are adopting. Most likely we can't prove because the > premise we choose to build on is abstract - at least has elements of > abstraction that won't allow a clear logical conclusion. > > The issue is, where will AGI get the assumptions? And, how rigorous will > the process be for accepting a new assumption? > > > > > [https://www.listbox.com/images/feed-icon-10x10.jpgecd5649.jpg?uri=aHR0cHM6Ly93d3cubGlzdGJveC5jb20vaW1hZ2VzL2ZlZWQtaWNvbi0xMHgxMC5qcGc]<https://www.listbox.com/member/archive/rss/303/26941503-0abb15dc> > > [https://www.listbox.com/images/listbox-logo-small.pngecd5649.png?uri=aHR0cHM6Ly93d3cubGlzdGJveC5jb20vaW1hZ2VzL2xpc3Rib3gtbG9nby1zbWFsbC5wbmc] > <http://www.listbox.com> > [https://www.listbox.com/images/feed-icon-10x10.jpgecd5649.jpg?uri=aHR0cHM6Ly93d3cubGlzdGJveC5jb20vaW1hZ2VzL2ZlZWQtaWNvbi0xMHgxMC5qcGc]<https://www.listbox.com/member/archive/rss/303/24379807-653794b5> > | > [https://www.listbox.com/images/listbox-logo-small.pngecd5649.png?uri=aHR0cHM6Ly93d3cubGlzdGJveC5jb20vaW1hZ2VzL2xpc3Rib3gtbG9nby1zbWFsbC5wbmc] > <http://www.listbox.com> > [https://www.listbox.com/images/feed-icon-10x10.jpgecd5649.jpg?uri=aHR0cHM6Ly93d3cubGlzdGJveC5jb20vaW1hZ2VzL2ZlZWQtaWNvbi0xMHgxMC5qcGc]<https://www.listbox.com/member/archive/rss/303/26941503-0abb15dc> > | > [https://www.listbox.com/images/listbox-logo-small.pngecd5649.png?uri=aHR0cHM6Ly93d3cubGlzdGJveC5jb20vaW1hZ2VzL2xpc3Rib3gtbG9nby1zbWFsbC5wbmc] > <http://www.listbox.com> > > > AGI | Archives<https://www.listbox.com/member/archive/303/=now> > [https://www.listbox.com/images/feed-icon-10x10.jpgecd5649.jpg?uri=aHR0cHM6Ly93d3cubGlzdGJveC5jb20vaW1hZ2VzL2ZlZWQtaWNvbi0xMHgxMC5qcGc] > <https://www.listbox.com/member/archive/rss/303/26941503-0abb15dc> | > Modify<https://www.listbox.com/member/?&> Your Subscription > [https://www.listbox.com/images/listbox-logo-small.pngecd5649.png?uri=aHR0cHM6Ly93d3cubGlzdGJveC5jb20vaW1hZ2VzL2xpc3Rib3gtbG9nby1zbWFsbC5wbmc] > <http://www.listbox.com> > > > > ------------------------------------------- > AGI > Archives: https://www.listbox.com/member/archive/303/=now > RSS Feed: https://www.listbox.com/member/archive/rss/303/11943661-d9279dae > Modify Your Subscription: > https://www.listbox.com/member/?& > Powered by Listbox: http://www.listbox.com > ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
