I will take a look at opencog. I am thinking of writing a simple
cross-categorical note book. I would like to start with something simple,
where the user creates the categories but then I would like to write an ai
part where the ai makes guesses about the categorization. But I want to
have be based on words used in sentences not just word occurrence rates. It
would not the same as basing it on general knowledge in some way.
Jim Bromer

On Mon, Feb 25, 2019 at 9:45 PM Linas Vepstas <[email protected]>
wrote:

> Hi Jim, OK thanks!  I don't know what you enjoy, what you are good at.  So
> I don't know what to recommend.  When you say "a simpler AI project", I
> don't know what that means. If you just want to learn -  I dunno - Buy
> books and study them? Create software and see what happens? Read and think
> and repeat and you'll learn something? If you want to help with a larger
> project like opencog - there's 1001 things that need to be done, technical
> and non-technical.We need cheerleaders and janitors, developers and
> researchers. You name it. Opencog is not a small and simple project, but
> its also a zillion times smaller and simpler than whatever facebook and IBM
> and google are doing.  But simply taking some inspired idea, and converting
> it into software, that's just a difficult task. No matter what the idea.
>
> --linas
>
> On Mon, Feb 25, 2019 at 3:10 PM Jim Bromer <[email protected]> wrote:
>
>> Thank you for that introduction to mathematical symbolic interpretation.
>> The notation was familiar enough that I could follow what you were saying
>> but the explanations helped me to avoid getting confused. I found it to be
>> very interesting and I am thinking about it. As a matter of fact, I want to
>> work on a simpler AI type of project which would make the insights that you
>> must have very useful to me. Unfortunately, I am not sure I could find the
>> time to be of much use to you. If you ever decide to write a textbook about
>> this topic, I would be very interested. If you do write any additional
>> chapters I think you should make sure that you include a few worked
>> examples. If there is anyway I could help you without working full time on
>> the project, let me know.
>> Jim Bromer
>>
>>
>> On Sun, Feb 24, 2019 at 11:46 PM Linas Vepstas <[email protected]>
>> wrote:
>>
>>> Attached is a PDF that reviews a relationship between symbols and
>>> reality. Depending on your outlook, you will probably find it to be either
>>> trivial, or wrong, or useless or all of the above. -- linas
>>>
>>> On Sun, Feb 24, 2019 at 2:32 PM Nanograte Knowledge Technologies <
>>> [email protected]> wrote:
>>>
>>>> You'll probably need both, not one vs the other. I'd think, if the same
>>>> soft architecture was used for a neural net as the symbol net, the symbol
>>>> net would eventually outperform the neural net, but only by virtue of the
>>>> data integrity of the neural net. The symbol net could be viewed as one
>>>> abstraction of a neural net. The symbol net could be integrated with
>>>> floating memory. This seems unlikely for a neural net to achieve on its
>>>> own.
>>>>
>>>> I do not have have all the "correct" terms here. I also try and stay
>>>> away from some terms on purpose, so as to not become trapped by them. To me
>>>> some of these ideas are functional concepts - a dichotomy.
>>>>
>>>> From a dynamical memory store it should be possible to deploy a
>>>> functional "RNA"-type coding schema as possibly the most-abstract
>>>> de-abstraction of the whole system. This is the domain of paradox I
>>>> referred to before. I'll call this paradox because time ceases to exist
>>>> somewhere along this point, before some of it, or all, or nothing,
>>>> fragment(s) into multiple versions of 1 entity. This is a point (or
>>>> universe) where an iterative spark of tacit knowing (read possible
>>>> consciousness) probably occurs.
>>>>
>>>> I think the previously-mentioned "location" may be an "area" from which
>>>> tacit knowledge might enter the timespace continuum before spinning off
>>>> explicit artifacts. This is because in my view, location has relevance for
>>>> gravitational force (if gravity would still exist as such in science after
>>>> another 10 years). I tend to agree how the process I described might be
>>>> pertinent to  an aspect of recombination, but one has to think
>>>> limitlessness. Recombination -as we seemingly know it biologically - seems
>>>> to be an insular concept. However, I think existentially it is a much
>>>> greater integral holistically than we may currently be able to imagine. At
>>>> this point, one would be given to much speculation.
>>>>
>>>> In my mind's eye, at a conceptual point of spark, I see an
>>>> environmental state of all and nothing and everything in between, finite
>>>> infinity and so on. Perhaps then, morphogenesis. I do not call this state
>>>> "life". At best, I'd refer to it as a superposition within informational
>>>> communication.
>>>>
>>>> I'll need to put my theory to the test though. Hopefully, in my
>>>> lifetime then. :-)
>>>>
>>>> Robert Benjamin
>>>>
>>>> ------------------------------
>>>> *From:* Jim Bromer <[email protected]>
>>>> *Sent:* Sunday, 24 February 2019 7:25 PM
>>>> *To:* AGI
>>>> *Subject:* Re: [agi] Some thoughts about Symbols and Symbol Nets
>>>>
>>>> Nano: I thought that you might be thinking of something like that. As I
>>>> tried to form a response I finally have started to make sense of it. A
>>>> discrete mathematical algorithm will suffer from the combinatorial
>>>> complications. A deep learning net seems to be able to deal with those
>>>> kinds of problems - as long as there are good approximate solutions
>>>> available. So then, is there anyway a symbol-net could outperform neural
>>>> nets (if rapidity was not a fundamental problem)? It seems to me that you
>>>> may be approaching  a possible solution to that kind of situation. Does
>>>> that analysis make sense to you?
>>>> Jim Bromer
>>>>
>>>>
>>>> On Sat, Feb 23, 2019 at 3:39 PM Nanograte Knowledge Technologies <
>>>> [email protected]> wrote:
>>>>
>>>> <https://mathinsight.org/definition/network>In other words...the
>>>> purpose should be functional RNA.
>>>>
>>>> Now, is an AGI blueprint justified?
>>>>
>>>> https://en.wikipedia.org/wiki/Gene_regulatory_network
>>>> <https://en.wikipedia.org/wiki/Gene_regulatory_network>
>>>> Gene regulatory network - Wikipedia
>>>> <https://en.wikipedia.org/wiki/Gene_regulatory_network>
>>>> A gene (or genetic) regulatory network (GRN) is a collection of
>>>> molecular regulators that interact with each other and with other
>>>> substances in the cell to govern the gene expression levels of mRNA and
>>>> proteins. These play a central role in morphogenesis, the creation of body
>>>> structures, which in turn is central to evolutionary developmental biology
>>>> (evo-devo).
>>>> en.wikipedia.org
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> ------------------------------
>>>> *From:* Jim Bromer <[email protected]>
>>>> *Sent:* Saturday, 23 February 2019 6:31 PM
>>>> *To:* AGI
>>>> *Subject:* Re: [agi] Some thoughts about Symbols and Symbol Nets
>>>>
>>>> So now I am beginning to wonder if nano was talking about some kind of
>>>> syntactic rules of combination for symbols and that all knowledge would
>>>> have to 'emerge' by selection of which combination, within a network, were
>>>> selected by learning.
>>>> Jim Bromer
>>>>
>>>>
>>>> On Fri, Feb 22, 2019 at 1:54 AM Nanograte Knowledge Technologies <
>>>> [email protected]> wrote:
>>>>
>>>> If this were the case, I'd agree with you. What I'm proposing is
>>>> content independent and context dependent. It is suitable for CAS
>>>> applications. It is not "designed"to be constrained, but to identify and
>>>> normalize a primary, contextual constraint in order to resolve it in an
>>>> adaptive sense. Meaning, humans do not resolve it, but the
>>>> contextually-bound instance of the system does. By implication, all
>>>> possible meanings of the symbol are always resident and latent. However,
>>>> the decisive meaning for a particular context is alive for the duration of
>>>> that contextual reference in the greater schema of information. In other
>>>> words, the correct answer is always possible within a particular context.
>>>> Such is the basis of critical thinking, to derive the correct answer to
>>>> every situation. Yes, there is an underlying assumption, which is that a
>>>> correct answer exists for every context, but this could be proven
>>>> scientifically.
>>>>
>>>> Previously, mention was made in the forum about hierarchy (meaning
>>>> control). Having hierarchy within a systems constructs provides a system
>>>> with robustness and integrity, which translates into informational
>>>> reliability. Now, it seems the question of validity has been settled, but
>>>> not the one on reliability. What I'm proposing already has embedded into
>>>> its language what could be termed validity and reliability, at scale.
>>>>
>>>> That is where the analogy of the tyre hitting the tar has relevance, or
>>>> the point in project and program management where the the essential truth
>>>> hits home. It is where the absolute impact on a situation has most effect.
>>>> We could also argue how it resembles the point of effective complexity,
>>>> which is the point-of-reasoning we are all desire within an AGI entity.
>>>>
>>>> You stated: "The term 'context-free' refers to the syntactic context,
>>>> not the greater global context (of variable type definitions or
>>>> redefinitions and so on)."
>>>>
>>>> >>I strongly disagree with this view. In a semantic system, which I
>>>> contend is required for a symbolic system to become operational, syntax
>>>> lends itself to context specificality. I think that point was born out via
>>>> recent discussions on the forum.
>>>>
>>>> I think no designer should (be allowed) arbitrarily decide local and
>>>> global boundaries. That's a recipe for disaster. Boundaries are outcomes of
>>>> the inherent (natural) design resident within components and collective
>>>> contexts. In addition to a specified, context boundary, the underlying
>>>> methodology should allow for scalability, which is not only an issue of
>>>> size, but also of adaptive scope (implying boundary adaptiveness). In this
>>>> sense, a contextual/systems boundary would be structured/unstructured in a
>>>> construct of thesis/antithesis - 2 parts of the same coin. Perhaps in using
>>>> this approach, we would achieve Haramein's et al's perspective on a
>>>> finite-infinity in a computational model.
>>>>
>>>> When looked at via a report, or a snapshot view, such a system would
>>>> appear to be structured (which it also is). However, if you could view it
>>>> as a continuous value stream, as a movie, it would be possible to watch
>>>> (and trace) how it morphed relative to its adaptive algorithm - as an
>>>> unstructured system. In time, for each specific context, it should become
>>>> possible to identify the patterns of such morphing, and apply predictive
>>>> algorithms.
>>>>
>>>> I think one outcome (there are multiple outcomes) of such a system
>>>> would resemble a Symbol Net. It should theoretically be possible to extract
>>>> such nets from the "live" system. I think this is rather similar to how we
>>>> do it within society today.
>>>>
>>>>  Robert Benjamin
>>>> ------------------------------
>>>> *From:* Jim Bromer <[email protected]>
>>>> *Sent:* Thursday, 21 February 2019 11:46 PM
>>>> *To:* AGI
>>>> *Subject:* Re: [agi] Some thoughts about Symbols and Symbol Nets
>>>>
>>>> A contextual reference framework, designed to limit the meaning of a
>>>> symbol to one meaning within a particular context, would only displace the
>>>> ambiguity - unless the language was artificially designed to be that way.
>>>> So called 'context-free' languages, ironically enough, do just that. They
>>>> have some value in AI, but it is difficult to see how it could be used as
>>>> an effective basis for stronger AI. The term 'context-free' refers to the
>>>> syntactic context, not the greater global context (of variable type
>>>> definitions or redefinitions and so on). Perhaps the term is misunderstood
>>>> or misused in compiler design, but, a lot like applied logic, its
>>>> application is useful because it can be limited to 'a framework' (like a
>>>> local function and so on). So perhaps industry did develop a way to limit
>>>> ambiguity within a contextual framework, but so far it has not proven to be
>>>> very useful in stronger AI. The nature of *limiting* ambiguity of a symbol
>>>> (or possible referential signification) does not seem to be a very powerful
>>>> tool to rely on when you are trying to stretch the reach of current (or 30
>>>> year old) ideas to attain greater powers of 'understanding'.
>>>> Jim Bromer
>>>>
>>>>
>>>> On Thu, Feb 21, 2019 at 2:49 PM Nanograte Knowledge Technologies <
>>>> [email protected]> wrote:
>>>>
>>>> If one had a contextual reference framework, each symbol would always
>>>> have one meaning within a particular context. Searches would always be
>>>> optimal. An example of this is evidenced within the Japanese language. So,
>>>> the 30+ years of waiting was for no good reason. If only the industry had
>>>> developed appropriate theory for dealing with scalable ambiguity, which it
>>>> probably had.
>>>>
>>>> ------------------------------
>>>> *From:* Jim Bromer <[email protected]>
>>>> *Sent:* Thursday, 21 February 2019 8:13 PM
>>>> *To:* AGI
>>>> *Subject:* Re: [agi] Some thoughts about Symbols and Symbol Nets
>>>>
>>>> I asked myself the question: If a theory of symbols was a feasible
>>>> basis for stronger AI, then the earlier efforts in discrete AI or weighted
>>>> reasoning should have show some promise. They should have worked. So why
>>>> didn't they work? Then I remembered that they did work with small data
>>>> sets. GOFAI did work as long as it could make a rapid search through the
>>>> possible candidates of meaning, but because combinations of symbols have
>>>> meaning, and because each symbol may have more than one meaning or referent
>>>> the problems of combinatorial complications presented a major obstacle to
>>>> developing the theories much further, My opinion is that the ambiguities or
>>>> multiple possible relevancies of a symbol (sub-net) can themselves be used
>>>> to narrow the possible meaning of the symbol (sub-net) when needed in
>>>> reasoning. We just need a huge amount of memory in order to create an index
>>>> of generalizations to use the information adequately. We now have that
>>>> scale of memory and processor speed available to us so we can try things
>>>> that could not be tried in the 1970s and 80s.
>>>> Jim Bromer
>>>>
>>>>
>>>> On Tue, Feb 19, 2019 at 12:45 AM Nanograte Knowledge Technologies <
>>>> [email protected]> wrote:
>>>>
>>>> Linas, Mike and Jim
>>>>
>>>> I find this to be a most-interesting conversation. Primarily, because
>>>> it suggests that the development of AGI may not only be challenged by the
>>>> development of competent theory, but also by programming capabilities to
>>>> put the theory into practice.
>>>>
>>>> Evolving such an architecture then, should desired outcomes be for an
>>>> AGI entity to achieve self-theory and self-programming? In its
>>>> most-simplistic from, a symbol is but a suitable abstraction of a greater
>>>> reality, similarly to how a symbol of a red-heart might be an abstraction
>>>> of a sentient being. Concept? Context? Meaning? Transaction.
>>>>
>>>> Who, or what decides what the symbolic world should look like and its
>>>> meaningfulness? The global state of social evolution may cause terrible
>>>> confusion in any learning entity. The learning objectives should be
>>>> specific, not generalized. Isn't learning incorrectly worse than not
>>>> learning at all?
>>>>
>>>> I think, there should be a general agi-architecture, replete with the
>>>> capacity to develop and function within a generic world view. Furthermore,
>>>> I think the real value would be derived from specialized AGI. Maybe beyond
>>>> that, an AGI architecture would - in future - morph via its own social
>>>> networking and inherent capabilities to become more than the sum of its
>>>> parts.
>>>>
>>>> To do so, would take a lot more than intersections. I agree with the
>>>> statements made about binary/vector theory, but it seems obvious to me that
>>>> this would not be sufficient for this task. You implied fractals. To my
>>>> mind, that would be the only way to proceed. As such, I think the primary
>>>> issue remains a design issue.
>>>>
>>>> Robert Benjamin
>>>>
>>>> ------------------------------
>>>> *From:* Linas Vepstas <[email protected]>
>>>> *Sent:* Monday, 18 February 2019 10:36 PM
>>>> *To:* AGI
>>>> *Subject:* Re: [agi] Some thoughts about Symbols and Symbol Nets
>>>>
>>>>
>>>>
>>>> On Mon, Feb 18, 2019 at 1:17 PM Mike Archbold <[email protected]>
>>>> wrote:
>>>>
>>>> I'm not sure I completely follow your point, but I sort of get it.
>>>>
>>>> I tend to think of symbols as one type of the "AI stuff" a computer
>>>> uses to think with -- the other main type of "AI stuff" being neural
>>>> networks. These have analogies to the "mind stuff" we use to think
>>>> with.
>>>>
>>>>
>>>> Symbol systems and neural-net systems can be seen to be variants of the
>>>> same thing; two sides of the same coin. I posted an earlier thread on this.
>>>> There's a 50-page long PDF with math, here:
>>>> https://github.com/opencog/opencog/raw/master/opencog/nlp/learn/learn-lang-diary/skippy.pdf
>>>>
>>>> roughly: both form networks. They differ primarily in how they
>>>> represent the networks, and how they assign weights to network connections
>>>> (and how they update weights on network connections).
>>>>
>>>>
>>>> On their own, symbols don't mean anything, of course, and inherently
>>>> don't contain "understanding" in any definition of understanding.
>>>>
>>>> Is there a broad theory of symbols? We kind of proceed with loose
>>>> definitions. I remember reading the Newell and Simon works, and they
>>>> say AI strictly in terms of symbols and LISP (as I recall anyway).
>>>>
>>>>
>>>> Yes. The "broad theory of symbols" is called "model theory" by
>>>> mathematicians. It's highly technical and arcane. It's most prominent
>>>> distinguishing feature as that everything is binary:  it is or it ain't.
>>>> Something is true, or false.  A formula takes values, or there is no such
>>>> formula. A relation binds two things together, or there is no relation.
>>>> There's no blurry middle-ground.
>>>>
>>>> So, conventionally, networks of symbols, and the relations between
>>>> them, and the formulas transforming them -- these form a network, a graph,
>>>> and everything on that network/graph is a zero or a one -- an edge exists
>>>> between two nodes, or it doesn't.
>>>>
>>>> The obvious generalization is to make these fractional, to assign
>>>> weights. Neural nets do this. But neural nets do something else, that they
>>>> probably should not: they jam everything into vectors (or tensors) This is
>>>> kind-of OK, because the algebra of a graph is a lot like the algebra of a
>>>> vector space, and the confusion between the two is an excusable mistake: it
>>>> takes some sophistication to realize that they are only similar, but not
>>>> the same.
>>>>
>>>> I claim: fix both these things, and you've got a winner.  Use symbolic
>>>> systems, but use fractional values, not 0/1 relations.  Find a good way of
>>>> updating the weights. So, deep-learning is a very effective weight-update
>>>> algorithm. But there are other ways of updating weights too (that are
>>>> probably just as good or better.  Next, clarify the
>>>> vector-space-vs-graph-algebra issue, and then you can clearly articulate
>>>> how to update weights on symbolic systems, as well.
>>>>
>>>> (Quickly explained: probabilities are not rotationally-symmetric under
>>>> the rotation group SO(N) whereas most neural-net vectors are: this is the
>>>> spot where deep-learning "gets it wrong": it incorrectly mixes gibbs
>>>> training functions with rotational symmetry.)
>>>>
>>>> So Jim is right: discarding symbolic systems in favor of neural nets is
>>>> a mistake; the path forward is at the intersection of the two: a net of
>>>> symbols, a net with weights, a net with gradient-descent properties, a net
>>>> with probabilities and probability update formulas.
>>>>
>>>> -- Linas
>>>>
>>>>
>>>> On 2/18/19, Jim Bromer <[email protected]> wrote:
>>>> > Since I realized that the discrete vs weighted arguments are passe I
>>>> > decided that thinking about symbol nets might be a better direction
>>>> for me,
>>>> >
>>>> > 1. A symbol may be an abstracted 'image' of a (relatively) lower level
>>>> > object or system.
>>>> >   An image may consist of a feature of the referent, it may be an
>>>> icon of
>>>> > the referent or it may be a compressed form of the referent.
>>>> > 2. A symbol may be more like a 'label' for some object or system.
>>>> > 3. A generalization may be represented as an image of what is being
>>>> > generalized but it also may be more of a label.
>>>> > 4. An 'image', as I am using the term, may be derived from a part or
>>>> > feature of an object or from a part of a system but it may be used to
>>>> refer
>>>> > to the object or system.
>>>> > 5. An image or label may be used to represent a greater system. A
>>>> system
>>>> > may take on different appearances from different vantage points, and
>>>> > analogously, some features of interest may be relevant in one context
>>>> but
>>>> > not from another context. A symbol may be correlated with some other
>>>> > 'object' and may stand as a referent to it in some contexts.
>>>> >
>>>> > So, while some symbols may be applied to or projected onto a 'lower'
>>>> corpus
>>>> > of data, others would need to use an image to project onto the data
>>>> field.
>>>> > I use the term, 'lower' somewhat ambiguously, because I think it is
>>>> useful
>>>> > to symbolize a system of symbols so a 'higher' abstraction of a system
>>>> > might also be used at the same level. And it seems that a label would
>>>> have
>>>> > to be associated with some images if it was to be projected against
>>>> the
>>>> > data.
>>>> >
>>>> > One other thing. This idea of projecting a symbol image onto some
>>>> data, in
>>>> > order to compare the image with some features of the data, seems like
>>>> it
>>>> > has fallen out of favor with the advancements of dlnns and other
>>>> kinds of
>>>> > neural nets. Projection seems like such a fundamental process that I
>>>> cannot
>>>> > see why it should be discarded just because it would be relatively
>>>> slow
>>>> > when used with symbol nets. And, there are exceptions, GPUs, for
>>>> example,
>>>> > love projecting one image onto another.
>>>> > Jim Bromer
>>>> 
>>>> 
>>>> --
>>>> cassette tapes - analog TV - film cameras - you
>>>> 
>>>
>>> --
>>> cassette tapes - analog TV - film cameras - you
>>>
>
> --
> cassette tapes - analog TV - film cameras - you
> *Artificial General Intelligence List <https://agi.topicbox.com/latest>*
> / AGI / see discussions <https://agi.topicbox.com/groups/agi> +
> participants <https://agi.topicbox.com/groups/agi/members> + delivery
> options <https://agi.topicbox.com/groups/agi/subscription> Permalink
> <https://agi.topicbox.com/groups/agi/Tcc0e554e7141c02f-Ma10490221fd36de4ab48363c>
>

------------------------------------------
Artificial General Intelligence List: AGI
Permalink: 
https://agi.topicbox.com/groups/agi/Tcc0e554e7141c02f-Ma26c9c0bdcb8cf33faef54be
Delivery options: https://agi.topicbox.com/groups/agi/subscription

Reply via email to