Hey Aaron, Why don't you join us in OpenCog ... Ruiting Lian is currently attempting to do exactly the same thing in OpenCog, and could use some help ;)
See the PDF at http://wiki.opencog.org/w/Link2Atom for a general description of the approach... ben On Mon, Aug 27, 2012 at 7:10 PM, Aaron Hosford <[email protected]> wrote: > That's the base assumption for my current project. I'm starting from human > language and attempting to derive an internal representation that > corresponds roughly to that used by humans. It is my hope that once an > appropriate model of how humans internally represent knowledge is > available, the actual mental computations we perform to handle higher-level > rational thought should become much more amenable to understanding and > analysis. This seems self evident to me, since clearly human beings have > solved the GI problem, which means we probably have some sort of internal > representation that makes the sorts of representational gymnastics that are > necessary for GI much simpler to perform. This approach has the advantage, > as well, that its accuracy as a model of human internal knowledge > representation should be directly verifiable in experiments with human > subjects. > > The idea is that I can run a parser on a piece of natural language, > extract out the relationships between the words as a semantic net, convert > that format into another semantic net that represents the meaning of the > language sample, and then reverse the flow back to natural language which > is identical in meaning but may be stated differently. When the meaning has > been extracted and represented using the internal format, it can be linked > up with other semantic nets that represent the meanings of other > statements/questions. This combined net in turn can be analyzed directly as > a collection of logical predicates and queries in which the binding of two > symbols (word/phrase occurrences) to a common referent are directly > represented as links from those symbols to that referent's node. New > statements/queries can be generated via inference rules and other daemons, > and then converted to natural language using the parser, etc. in reverse. > > I have already built a small system as a proof of concept with Boolean > links -- either a semantic link exists or doesn't, rather than allowing > links to have real-valued strengths -- and was able to resolve anaphora > moderately well, given its toy nature. Since this initial implementation > left me unsatisfied with how uncertainty was handled, I'm working now on > rebuilding the system using real-valued links that represent > probability/uncertainty, similarly to the truth values used by the term > logic-based inference system of NARS ( > https://sites.google.com/site/narswang/). Adding in the ability to > represent uncertainty will allow the system to compare alternatives and > choose the most salient anaphoric referent *that makes sense*, taking > advantage of knowledge the system has already acquired to determine what > makes sense, rather than just taking the most salient/obvious choice in > terms of raw language structure independent of conceptual context. > > > > On Mon, Aug 27, 2012 at 4:55 PM, Anastasios Tsiolakidis < > [email protected]> wrote: > >> As I started reading I thought to myself "I told you 1000 times, it >> depends on the criteria". Reading on, I saw that it is precisely the >> criteria you use as a parameter. Well, I'd like to find out the >> programming language that makes the most money while giving >> immortality :) A little more seriously, if the criteria are cognitive, >> as they often are in the real world, you'd be digging yourself a hole >> too deep to get out of. On the other hand, if the criteria are >> domain-specific, relating to well-behaved domains, I am afraid we are >> heading towards tautologies and trivialities. Something like >> Mathematica would be optimal for algebra, analysis, gravity, mechanics >> etc (though what about instead of calculating a parachute drop >> actually measure a real parachute drop), for economics, psychology, >> necromancy most things would do equally badly, and for AGI all options >> have so far being worse than bad. Mind you, I am in the process of >> defining an AGI architecture not as a compression problem but as a >> distributed computation problem, and I would challenge you to answer >> the question: >> >> Which programming language/mechanism would be ideal for calculating X >> as quickly as possible. >> >> where X, for the sake of argument, is just a/any "heavy calculation" >> without necessarily any of the anomalies of chaotic behavior, pi's >> infinite series etc. It is not that I expect intelligence to arise out >> of PDEs and integrals, rather I am asking which is the "perfect" >> distributed system for calculus, as I am expecting your answer to take >> the form of multipliers and other exotic units all converging in an >> addition pipeline. I still can't help thinking that the fastest way >> for parallel computations is the actual experiment, after all we have >> the 3/n body problem and a ton of mathematics OR just an experiment >> with n bodies in a field. >> >> With regards to a possible language for AGI, I don't see how you can >> do much better than a human language. Never mind Turing completeness, >> we have GI completeness here (except for that part of human language, >> perhaps 100% of it, that gets its meaning from its grounding, its >> grounding from its embodiment, and its embodiment from - god?) >> >> AT >> >> On Mon, Aug 27, 2012 at 10:44 PM, Russell Wallace >> <[email protected]> wrote: >> > On Mon, Aug 27, 2012 at 9:12 PM, Ben Goertzel <[email protected]> wrote: >> >> For domains in which one is concerned with recognizing large ensembles >> >> of weak patterns, the language one uses to represent patterns can make >> >> a big difference... >> >> >> >> Image analysis, genetic data analysis and financial prediction are >> >> contexts in which I've found this to be the case >> >> >> >> In these settings, if one does pattern recognition via automated >> >> program learning with an Occam bias, >> >> the underlying language relative to which the Occam bias is expressed >> >> makes a big difference... >> > >> > Absolutely, but these overheads are not constants - the computational >> > cost of a poor choice of representation language is typically >> > exponential. >> > >> >> From a different direction, consider Hutter's proof that AIXI-tl is as >> >> good as any other reinforcement learning system ... up to an arbitrary >> >> constant. >> > >> > Well, much violence is being done to the word 'constant' in this case. >> > Sure, f(N) is constant for a given N, but... :) >> > >> > >> > ------------------------------------------- >> > AGI >> > Archives: https://www.listbox.com/member/archive/303/=now >> > RSS Feed: >> https://www.listbox.com/member/archive/rss/303/14050631-7d925eb1 >> > Modify Your Subscription: https://www.listbox.com/member/?& >> > Powered by Listbox: http://www.listbox.com >> >> >> ------------------------------------------- >> AGI >> Archives: https://www.listbox.com/member/archive/303/=now >> RSS Feed: >> https://www.listbox.com/member/archive/rss/303/23050605-bcb45fb4 >> >> Modify Your Subscription: https://www.listbox.com/member/?& >> Powered by Listbox: http://www.listbox.com >> > > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/212726-11ac2389> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com> > -- Ben Goertzel, PhD http://goertzel.org "My humanity is a constant self-overcoming" -- Friedrich Nietzsche ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-c97d2393 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-2484a968 Powered by Listbox: http://www.listbox.com
