I believe my response to Hawkins' book, written in 2004, still holds as a response to Kurzweil's book (with its similar theme) as well:
http://www.goertzel.org/dynapsyc/2004/OnBiologicalAndDigitalIntelligence.htm -- Ben Goertzel On Mon, Nov 19, 2012 at 4:42 PM, Micah Blumberg <[email protected]> wrote: > YKY your approach does not personally make sense to me. Naturally I hope > you are successful. Have you read either "On Intelligence" by Jeff Hawkins > or "How to create a mind" by Ray Kurzweil. The latter book was recently > published. I suspect that many people trying to build AGI will be able to > consider a new approach to after reading it. Please let me know if you are > aware of that kind of AI and what advantages your planned system may have > over it. > > > On Sun, Nov 18, 2012 at 9:20 PM, Aaron Hosford <[email protected]>wrote: > >> My matrix representation is based on the observation that AB != BA in the >>> composition of concepts, for example "john loves mary != mary loves john". >>> I'm still trying to determine if the matrix representation is sound and >>> consistent with our common-sense view of concepts, but it looks promising... >> >> >> I may be jumping to conclusions, but it sounds like what you're doing is >> placing A/"john", B/"mary", etc. as labels for both rows and columns, where >> the row label represent the first value and the column represents the >> second (or maybe vice versa), so that you can represent "john loves mary" >> by placing a nonzero value in entry indexed by row "john" and column >> "mary". The most obvious interpretation for this nonzero value would be the >> probability that "john loves mary" is true, or it could possibly be >> certainty or some other similar measure. >> >> As you pointed out, this is a nice generalization of a basic adjacency >> matrix, with nonzero entries marking adjacencies (directed links or edges) >> in the conceptual graph. Matrices of this form could be mapped to directed >> graphs with links not only labeled "loves", but having a secondary label >> indicating the associated probability or certainty value of the link. In my >> own system, I keep my internal representation directly in terms of nodes & >> links rather than such a generalized adjacency matrix, but the specific >> data structure is really only an implementation detail, not intrinsic to >> the structure of the information being stored. We are basically doing the >> same thing, in other words, provided my understanding of your >> representational method is correct. >> >> I'm still trying to determine if the matrix representation is sound and >>> consistent with our common-sense view of concepts, but it looks promising... >> >> >> I have spent a lot of time thinking about this problem, albeit in terms >> of semantic nets rather than the matrices that correspond to them. I think >> your approach is consistent with how we internally work with concepts, but >> it is probably incomplete. For example, what happens when you want to >> represent the probability or certainty that "john" is the one participating >> as the subject in this "loves" relationship independently from the >> probability/certainty that "mary" is the direct object? Or what if you have >> lower probability/certainty for the classification of the relationship as >> "loves" (as opposed to say, "hates") than you do for the individuals >> filling the subject and direct object roles? It seems now that you need 3 >> probability/certainty numbers for each position in the matrix, one for >> subject, one for direct object, and one for relationship type. >> >> Instead of multiple values in a single cell of the matrix -- or rather 3 >> parallel matrices with the same row and column labels -- it seems more >> natural to me to make the clause/predicate/act/relationship a separate >> value in itself, so the original matrix is decomposed into (1) a "subject" >> matrix, mapping the probability/certainty of "john" (among other people) >> participating in the clause as the subject, (2) a "direct object" matrix, >> mapping the probability/certainty of "mary" (among other people) >> participating in the clause as the direct object, and (3) a "predicate" >> matrix, mapping the probability/certainty that the clause's predicate is >> "loves" (among other possible predicates). Each of these different >> matrices, "subject", "direct object", and "predicate", maps to a link label >> in the corresponding directed graph, and each link receives its own >> probability/certainty value independent of the others. The representational >> advantage of making the clause itself a row/column label only grows when >> prepositional phrases, adverbs, indirect objects, and other clausal >> modifiers start to come into play, and the utility of your matrix trick is >> preserved by simply adding a new matrix for each constituent role in the >> clause. >> >> >> >> On Sun, Nov 18, 2012 at 8:10 PM, YKY (Yan King Yin, 甄景贤) < >> [email protected]> wrote: >> >>> On Sat, Nov 10, 2012 at 3:44 AM, Aaron Hosford <[email protected]>wrote: >>> >>>> My "matrix trick" may be able to map propositions to a high-dimensional >>>>> vector space, but my assumption that concepts are matrices (that are also >>>>> rotations in space) may be unjustified. I need to find a set of criteria >>>>> for matrices to represent concepts faithfully. This direction is still >>>>> hopeful. >>>> >>>> >>>> I'm curious what your "matrix trick" is. Are you familiar with >>>> adjacency matrices? They're the simplest and most common way of >>>> representing directed graphs as matrices. >>>> http://en.wikipedia.org/wiki/Adjacency_matrix Semantic nets typically >>>> have sparse adjacency matrices, and there are a lot of good algorithms and >>>> libraries out there for efficiently representing and manipulating sparse >>>> matrices. >>>> >>> >>> >>> Sorry, almost missed your post. My method is to *represent* logical >>> atoms as square matrices with real entries. This allows me to convert the >>> matrices to vectors (by simply flattening the matrices) and calculate the >>> distances (ie similarities) between them via the Euclidean norm. >>> >>> My matrices are different from adjacency matrices which has integer >>> entries. The continuous values potentially allow learning via propagation >>> of errors similar to back-propagation for neural nets. >>> >>> My matrix representation is based on the observation that AB != BA in >>> the composition of concepts, for example "john loves mary != mary loves >>> john". I'm still trying to determine if the matrix representation is sound >>> and consistent with our common-sense view of concepts, but it looks >>> promising... >>> >>> YKY >>> *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> >>> <https://www.listbox.com/member/archive/rss/303/23050605-bcb45fb4> | >>> Modify <https://www.listbox.com/member/?&> Your Subscription >>> <http://www.listbox.com> >>> >> >> *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> >> <https://www.listbox.com/member/archive/rss/303/23601407-ccf7ca1d> | >> Modify <https://www.listbox.com/member/?&> Your Subscription >> <http://www.listbox.com> >> > > > > -- > > ~~ > Warmly, > > > Micah > 7 1 4 ) 6 9 9 - 4 2 1 3 (voicemail and texting same digits) > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/212726-c2d57280> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com> > -- Ben Goertzel, PhD http://goertzel.org "My humanity is a constant self-overcoming" -- Friedrich Nietzsche ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-c97d2393 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-2484a968 Powered by Listbox: http://www.listbox.com
