https://www.researchgate.net/publication/4363016_Last-mile_knowledge_engineering_Quest_for_the_holy_grail [https://www.researchgate.net/images/template/default_publication_preview_large.png]<https://www.researchgate.net/publication/4363016_Last-mile_knowledge_engineering_Quest_for_the_holy_grail>
Last-mile knowledge engineering: Quest for the holy grail?<https://www.researchgate.net/publication/4363016_Last-mile_knowledge_engineering_Quest_for_the_holy_grail> Download citation | Last-mile knowledge... | The problem of reliably structuring unseen knowledge, at scale, persists within systems engineering. An emergence-based method was developed to test the theory of applying de-abstraction reasoning to tacit-knowledge engineering. www.researchgate.net ________________________________ From: Jim Bromer via AGI <[email protected]> Sent: Monday, 25 June 2018 10:32 PM To: AGI Subject: Re: [agi] Discrete Methods are Not the Same as Logic Please provide a link to the method you are talking about. Jim Bromer On Sun, Jun 24, 2018 at 6:59 AM, Nanograte Knowledge Technologies via AGI <[email protected]> wrote: > Convention is a language. The program has to find a way to understand what > people say, and not say. It has to be able to learn the deeper meaning > within the human conversation, and systemically get to the heart of every > matter. It has to do this in the most-true manner, utilizing evidence-based > objectivity where possible. That method exists. It's the one I shared here. > That's exactly what it does. What it needs to become automated is the > development of its own GUI, as translator. > ________________________________ > From: Jim Bromer via AGI <[email protected]> > Sent: Saturday, 23 June 2018 3:55 PM > > To: AGI > Subject: Re: [agi] Discrete Methods are Not the Same as Logic > > I am thinking of a program that would learn by communicating with > people using language. It would learn from interacting with people. > The problem with that strategy is that it would tend to acquire > superficial knowledge. It would however be required to do some true > learning. One reason is that a person cannot think of all the > relations and implicit categories that an intelligent entity would > have to rely on. Secomdly, we cannot, at this time, understand all the > sorts of a knowledge items that it would need to gain greater > understanding. > It would not be given predetermined categories other than some default > second level abstract categories. These second level abstractions > might be concerned with abstractions of relations found in discrete > relationships that would be expected to found in networks of related > information. It would have to work around the complexities that might > develop. I am not talking about pure logic but discrete learning so > the np problem is not a problem. The "discrete networks" would also > include weighted reasoning of course. I am just saying that weighted > reasoning isn't necessary but that discrete learning, learning by > using ideas and developing principles of thought is important. > But I have to be able to develop this as an extremely simple > programming project that will quickly show some simple results (like > feasibility tests) or else I am not going to have anything to start > with. > > Jim Bromer > > > On Sat, Jun 23, 2018 at 8:51 AM, Nanograte Knowledge Technologies via > AGI <[email protected]> wrote: >> Jim >> >> I agree with making things simple, but one should not make it more simple >> than necessary. Any algorithm relying on deabstraction to provide proof of >> true learning would be highly complex. There's no simple solution to that >> problem. However, I'm enjoying your sentiment how, within deabstraction, >> even complexity should become relative over time. Maybe one day, the >> machine >> would've learned how to invent deabstraction algorithms until it became a >> simple matter of instinct. >> >> Since when do human beings discover all its learning by itself? That's a >> fallacy. An AGI platform also does not have to discover all of its >> learning >> by itself. It can be taught until such time it can learn how to organize >> resources in order to teach itself and learn via reflection. >> >> Rob >> ________________________________ >> From: Jim Bromer via AGI <[email protected]> >> Sent: Saturday, 23 June 2018 2:41 AM >> >> To: AGI >> Subject: Re: [agi] Discrete Methods are Not the Same as Logic >> >> Maybe I should use a name different than judgement. Reflection? >> Insightful reflection. The depth of the insight would be relative to >> how much knowledge, related to the questions being examined, was >> available. So in the primitive model this insight would not be very >> good and the program would have to be dependent on what the teacher >> could convey to it. But insight would have to be based on putting >> different kinds of information together. Novel insight might be >> reinforced simply by being in the ballpark, it would not have to be >> perfect as long as it was tagging along somewhere within the subject >> matter being discussed, described or within the boundaries of >> understanding something about a situation that was occurring. I think >> different agi's would have to be different if they were thinking for >> themselves - to some extent. >> Jim Bromer >> >> >> On Fri, Jun 22, 2018 at 3:10 PM, Mike Archbold via AGI >> <[email protected]> wrote: >>> Judgments are fascinating. It seems like most approaches are some >>> variation of reinforcement learning. What have you got in mind? One >>> thought from Hegel which always sticks in my mind is that a "judgment >>> could be other than what it is." So just think about that last >>> sentence. How on earth could anyone automate that? But, more so, two >>> distinct AGI's would always be different on that account. >>> >>> On 6/22/18, Jim Bromer via AGI <[email protected]> wrote: >>>> I need to start with something that is extremely simple and which will >>>> produce some kind of result pretty quickly. I have had various ideas >>>> about it for some time but what I see now is that a necessary >>>> advancement for AI would have to exhibit some kind of judgment about >>>> what it learns about. I realized the importance of making a program >>>> that could learn new ways of thinking. Since I believe that >>>> categorical reasoning is important then that means that it would not >>>> only have to use abstractions but it would also have to be able to >>>> discover abstractions of its own. This does not seem too difficult >>>> because I am not being unreasonable about requiring it to be a >>>> historical singularity inflexion point. I need to start with >>>> something simple that demonstrates an ability for true learning. What >>>> I see now is that it also has to exhibit some kind of simple >>>> judgement. I need to come up with simple judgement algorithms. I >>>> cannot get started unless I can come up with simple feasible models >>>> that I can test. >>>> I respectfully disagree with you about one thing. The elaboration of >>>> an extensive framework and management system is, in my opinion, a >>>> waste of time. It is like planning an AI program that will create AGI >>>> for you completely on its own. It might be ok to think about such a >>>> thing but it is nowhere to start out for an actual programming >>>> project. I have to start with something that is very simple and which >>>> can show some immediate results. For me, simplification is a necessity >>>> but it is also necessary to avoid the wrong kinds of simplification. >>>> Jim Bromer >>>> >>>> >>>> On Fri, Jun 22, 2018 at 12:13 AM, Nanograte Knowledge Technologies via >>>> AGI <[email protected]> wrote: >>>>> Jim, I think for this kind of reasoning to evolve, one would always >>>>> have >>>>> to >>>>> return to an ontological platform. For example, for reasoning, one >>>>> would >>>>> require a meta-methodology for reasoning effectively with. For >>>>> selectively >>>>> forgetting and learning, an evolution-based methodology is required. >>>>> For >>>>> managing Logic, one would need a suitable framework and management >>>>> system, >>>>> and so on. These are all critical components, or nodes, that would have >>>>> to >>>>> exist for self-optimized reasoning functionality to become >>>>> spontaneous.The >>>>> real IP lie not only in the methods, in the sense of AI apps. >>>>> >>>>> Yuu stated: "...DL story is compelling it is not paying out to stronger >>>>> AI >>>>> (Near AGI)..." >>>>>>>>Is it possible that AGI is an outcome, an act of becoming, and not a >>>>>>>> discrete objective at all? >>>>> >>>>> Rob >>>>> ________________________________ >>>>> From: Jim Bromer via AGI <[email protected]> >>>>> Sent: Thursday, 21 June 2018 5:20 PM >>>>> To: AGI >>>>> Subject: Re: [agi] Discrete Methods are Not the Same as Logic >>>>> >>>>> Symbol Based Reasoning is discrete, but a computer can use discrete >>>>> data that would not make sense to us so the term symbolic might be >>>>> misleading. I am not opposed to weighted reasoning (like neural >>>>> networks or Bayesian Networks) and I think reasoning has to use >>>>> networks of relations. If weighted networks can be thought of as a >>>>> symbolic network then that suggests that symbols may not be discrete >>>>> (as different from Neural Networks.) I just think that there is >>>>> something missing with DL, and while the Hinton...DL story is >>>>> compelling it is not paying out to stronger AI (Near AGI). For >>>>> example, I think that symbolic reasoning which is able to change its >>>>> categorical bases of reasoning is something that is badly lacking in >>>>> Discrete Learning. You don't want your program to forget everything it >>>>> has learned just because some doofus tells it to, and you do not want >>>>> it to write over the most effective methods it uses to learn just to >>>>> deal with some new method of learning. So, that, in my opinion is >>>>> where the secret may have been hiding. A program that is capable of >>>>> learning something new must be capable of losing its more primitive >>>>> learning techniques without wiping out the good stuff that it had >>>>> previously acquired. This requires some working wisdom. >>>>> I have been thinking about these ideas for a long time but now I feel >>>>> that I have a better understanding of how this insight might be used >>>>> to point to simple jumping off point. >>>>> Jim Bromer >>>>> >>>>> >>>>> On Thu, Jun 21, 2018 at 2:48 AM, Mike Archbold via AGI >>>>> <[email protected]> wrote: >>>>>> So, by "discrete reasoning" I think you kind of mean more or less "not >>>>>> neural networks" or I think some people say, or used to say NOT "soft >>>>>> computing" to mean, oh hell!, we aren't really sure how it works, or >>>>>> we can't create what looks like a clear, more or less deterministic >>>>>> program like in the old days etc.... Really, the challenge a lot of >>>>>> people, myself included, have taken up is how to fuse discrete (I >>>>>> simply call it "symbolic", although nn have symbols, typically you >>>>>> don't see them except as input and output) and DL which is such a good >>>>>> way to approach combinatorial explosion. >>>>>> >>>>>> To me reasoning is mostly conscious, and kind of like the way an >>>>>> expert system chains, logically. The understanding is something else >>>>>> riding kind of below it and less conscious but it has all the common >>>>>> sense rules of reality which constrain the upper level reasoning which >>>>>> I think is logical, like "if car won't start battery is dead" would be >>>>>> the conscious part but the understanding would include such mundane >>>>>> details as "a car has one battery" and "you can see the car but it is >>>>>> in space which is not the same thing as you" and "if you turn around >>>>>> to look at the battery the car is still there" and all such details >>>>>> which lead to an understanding. But understanding is an incredibly >>>>>> tough thing to make a science out of, although I see papers lately and >>>>>> conference topics on it. >>>>>> >>>>>> On 6/20/18, Jim Bromer via AGI <[email protected]> wrote: >>>>>>> I was just reading something about the strong disconnect between our >>>>>>> actions and our thoughts about the principles and reasons we use to >>>>>>> describe why we react the way we do. This may be so, but this does >>>>>>> not >>>>>>> show >>>>>>> how we come to understand basic ideas about the world. This attempt >>>>>>> to >>>>>>> make >>>>>>> a nearly total disconnect between reasons and our actual reactions >>>>>>> misses >>>>>>> something when it comes to explaining how we know anything, including >>>>>>> how >>>>>>> we learn to make decisions about something. One way to get around >>>>>>> this >>>>>>> problem is to say that it all takes place in neural networks which >>>>>>> are >>>>>>> not >>>>>>> open to insight about the details. But there is another explanation >>>>>>> which >>>>>>> credits discrete reasoning with the ability to provide insight and >>>>>>> direction and that is we are not able to consciously analyze all the >>>>>>> different events that are occurring at a moment and so we probably >>>>>>> are >>>>>>> reacting to many different events which we could discuss as discrete >>>>>>> events >>>>>>> if we had the luxury to have them all brought to our conscious >>>>>>> attention. >>>>>>> So logic and personal principles are ideals which we can use to >>>>>>> examine >>>>>>> our >>>>>>> reactions - and our insights - about the what is going on around us >>>>>>> but >>>>>>> it >>>>>>> is unlikely that we can catalogue all the events that surround us and >>>>>>> (partly) cause us to react the way we do. >>>>>>> >>>>>>> Jim Bromer >>>>>>> >>>>>>> On Wed, Jun 20, 2018 at 6:06 AM, Nanograte Knowledge Technologies via >>>>>>> AGI >>>>>>> < >>>>>>> [email protected]> wrote: >>>>>>> >>>>>>>> "As Julian Jaynes put it in his iconic book *The Origin of >>>>>>>> Consciousness >>>>>>>> in the Breakdown of the Bicameral Mind* >>>>>>>> >>>>>>>> Reasoning and logic are to each other as health is to medicine, or — >>>>>>>> better — as conduct is to morality. Reasoning refers to a gamut of >>>>>>>> natural >>>>>>>> thought processes in the everyday world. Logic is how we ought to >>>>>>>> think >>>>>>>> if >>>>>>>> objective truth is our goal — and the everyday world is very little >>>>>>>> concerned with objective truth. Logic is the science of the >>>>>>>> justification >>>>>>>> of conclusions we have reached by natural reasoning. My point here >>>>>>>> is >>>>>>>> that, >>>>>>>> for such natural reasoning to occur, consciousness is not necessary. >>>>>>>> The >>>>>>>> very reason we need logic at all is because most reasoning is not >>>>>>>> conscious >>>>>>>> at all." >>>>>>>> >>>>>>>> >>>>>>>> https://cameroncounts.wordpress.com/2010/01/03/mathematics-and-logic/ >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> <https://cameroncounts.wordpress.com/2010/01/03/mathematics-and-logic/> >>>>>>>> Mathematics and logic | Peter Cameron's Blog >>>>>>>> >>>>>>>> >>>>>>>> <https://cameroncounts.wordpress.com/2010/01/03/mathematics-and-logic/> >>>>>>>> Apologies: this will be a long post, and there will be more to come. >>>>>>>> But >>>>>>>> it may be useful to you if you are getting to grips with logic: I >>>>>>>> have >>>>>>>> tried to keep the overall picture in view. >>>>>>>> cameroncounts.wordpress.com >>>>>>>> >>>>>>>> >>>>>>>> ------------------------------ >>>>>>>> *From:* Jim Bromer via AGI <[email protected]> >>>>>>>> *Sent:* Wednesday, 20 June 2018 12:01 PM >>>>>>>> *To:* AGI >>>>>>>> *Subject:* Re: [agi] Discrete Methods are Not the Same as Logic >>>>>>>> >>>>>>>> Discrete statements are used in programming languages. So a symbol >>>>>>>> (a >>>>>>>> symbol phrase or sentence) can be used to represent both data and >>>>>>>> programming actions. Discrete Reasoning might be compared to >>>>>>>> something >>>>>>>> that has the potential to be more like an algorithm. (Of course, >>>>>>>> operational statements may be retained as data which can be run when >>>>>>>> needed) >>>>>>>> For an example of the value of Discrete Methods, let's suppose >>>>>>>> someone >>>>>>>> wanted more control over a neural network. Trying to look for logic >>>>>>>> in >>>>>>>> a neural network does not really make all that much sense if you >>>>>>>> want >>>>>>>> to find relationships between actions on the net and output. Using >>>>>>>> Discrete Methods makes a lot of sense. You might want to try >>>>>>>> fiddling >>>>>>>> with the weights of some of the nodes as the nn is running. If >>>>>>>> certain >>>>>>>> effects can be described (or sensed by some algorithm) then >>>>>>>> describing >>>>>>>> what was done and what effects were observed would be the next step >>>>>>>> in >>>>>>>> the research. Researchers are not usually able to start with >>>>>>>> detailed >>>>>>>> knowledge of exactly what is going on. So they need to start with >>>>>>>> descriptions of some actions they took and of what effects were >>>>>>>> observed. If these actions and effects can be categorized in some >>>>>>>> way >>>>>>>> then the chance that more effective observations will be obtained >>>>>>>> will >>>>>>>> increase. >>>>>>>> Jim Bromer >>>>>>>> >>>>>>>> >>>>>>>> On Tue, Jun 19, 2018 at 11:12 PM, Mike Archbold via AGI >>>>>>>> <[email protected]> wrote: >>>>>>>> > It sounds like you need both for AI, certainly there is always a >>>>>>>> > place >>>>>>>> > for logic. What's "discrete reasoning"? >>>>>>>> > >>>>>>>> > On 6/18/18, Jim Bromer via AGI <[email protected]> wrote: >>>>>>>> >> I am wondering about how Discrete Reasoning is different than >>>>>>>> >> Logic. >>>>>>>> >> I >>>>>>>> >> assume that Discrete Reasoning could be described, modelled or >>>>>>>> >> represented by Logic, but as a more practical method, logic would >>>>>>>> >> be >>>>>>>> >> a >>>>>>>> >> tool to use with Discrete Reasoning rather than as a >>>>>>>> >> representational >>>>>>>> >> substrate. >>>>>>>> >> >>>>>>>> >> Discrete Reasons and Discrete Reasoning can have meaning over and >>>>>>>> >> above the True False values of Logic (and the True False >>>>>>>> >> Relationships >>>>>>>> >> between combinations of Propositions.) >>>>>>>> >> >>>>>>>> >> Discrete Reasoning can have combinations that do not have a >>>>>>>> >> meaning >>>>>>>> >> or >>>>>>>> >> which do not have a clear meaning. This is one of the most >>>>>>>> >> important >>>>>>>> >> distinctions. >>>>>>>> >> >>>>>>>> >> It can be used in various combinations of hierarchies and/or in >>>>>>>> >> non-hierarchies. >>>>>>>> >> >>>>>>>> >> It can, for the most part, be used more freely with other >>>>>>>> >> modelling >>>>>>>> >> methods. >>>>>>>> >> >>>>>>>> >> Discrete Reasoning may be Context Sensitive in ways that produce >>>>>>>> >> ambiguities, both useful and confusing. >>>>>>>> >> >>>>>>>> >> Discrete Reasoning can be Active. So a statement about some >>>>>>>> >> subject >>>>>>>> >> might, for one example, suggest that you should change your >>>>>>>> >> thinking >>>>>>>> >> about (or representation of) the subject in a way that goes >>>>>>>> >> beyond >>>>>>>> >> some explicit propositional description about some object. >>>>>>>> >> >>>>>>>> >> You may be able to show that Logic can be used in a way to allow >>>>>>>> >> for >>>>>>>> >> all these effects, but I believe that there is a strong argument >>>>>>>> >> for >>>>>>>> >> focusing on Discrete Reasoning, as opposed to Logic, when you are >>>>>>>> >> working directly on AI. >>>>>>>> >> >>>>>>>> >> Jim Bromer >>>>>>>> *Artificial General Intelligence List >>>>>>>> <https://agi.topicbox.com/latest>* >>>>>>>> / AGI / see discussions <https://agi.topicbox.com/groups/agi> + >>>>>>>> participants <https://agi.topicbox.com/groups/agi/members> + >>>>>>>> delivery >>>>>>>> options <https://agi.topicbox.com/groups> Permalink >>>>>>>> >>>>>>>> >>>>>>>> >>>>>>>> <https://agi.topicbox.com/groups/agi/Tcc2adcdd20e1add4-M155d4762ea9c7b0f14fefd47> >>>>> Artificial General Intelligence List / AGI / see discussions + >>>>> participants >>>>> + delivery options Permalink >> Artificial General Intelligence List / AGI / see discussions + >> participants >> + delivery options Permalink > Artificial General Intelligence List / AGI / see discussions + participants > + delivery options Permalink ------------------------------------------ Artificial General Intelligence List: AGI Permalink: https://agi.topicbox.com/groups/agi/Tcc2adcdd20e1add4-M2cb6fbb9cc9ea093627d9334 Delivery options: https://agi.topicbox.com/groups
