Re: [agi] Mushed Up Decision Processes
2008/11/30 Ben Goertzel [EMAIL PROTECTED]: Could you give me a little more detail about your thoughts on this? Do you think the problem of increasing uncomputableness of complicated complexity is the common thread found in all of the interesting, useful but unscalable methods of AI? Jim Bromer Well, I think that dealing with combinatorial explosions is, in general, the great unsolved problem of AI. I think the opencog prime design can solve it, but this isn't proved yet... Good luck with that! In general, the standard AI methods can't handle pattern recognition problems requiring finding complex interdependencies among multiple variables that are obscured among scads of other variables The human mind seems to do this via building up intuition via drawing analogies among multiple problems it confronts during its history. Yes, so that people learn one problem, then it helps them to learn other similar ones. Is there any AI software that does this? I'm not aware of any. I have proposed a problem domain called function predictor whose purpose is to allow an AI to learn across problem sub-domains, carrying its learning from one domain to another. (See http://www.includipedia.com/wiki/User:Cabalamat/Function_predictor ) I also think it would be useful if there was a regular (maybe annual) competition in the function predictor domain (or some similar domain). A bit like the Loebner Prize, except that it would be more useful to the advancement of AI, since the Loebner prize is silly. -- Philip Hunt, [EMAIL PROTECTED] Please avoid sending me Word or PowerPoint attachments. See http://www.gnu.org/philosophy/no-word-attachments.html --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244id_secret=120640061-aded06 Powered by Listbox: http://www.listbox.com
Re: [agi] Mushed Up Decision Processes
Hi, I have proposed a problem domain called function predictor whose purpose is to allow an AI to learn across problem sub-domains, carrying its learning from one domain to another. (See http://www.includipedia.com/wiki/User:Cabalamat/Function_predictor ) I also think it would be useful if there was a regular (maybe annual) competition in the function predictor domain (or some similar domain). A bit like the Loebner Prize, except that it would be more useful to the advancement of AI, since the Loebner prize is silly. -- Philip Hunt, [EMAIL PROTECTED] How does that differ from what is generally called transfer learning ? ben g --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244id_secret=120640061-aded06 Powered by Listbox: http://www.listbox.com
Re: [agi] Mushed Up Decision Processes
2008/11/30 Ben Goertzel [EMAIL PROTECTED]: Hi, I have proposed a problem domain called function predictor whose purpose is to allow an AI to learn across problem sub-domains, carrying its learning from one domain to another. (See http://www.includipedia.com/wiki/User:Cabalamat/Function_predictor ) I also think it would be useful if there was a regular (maybe annual) competition in the function predictor domain (or some similar domain). A bit like the Loebner Prize, except that it would be more useful to the advancement of AI, since the Loebner prize is silly. -- Philip Hunt, [EMAIL PROTECTED] How does that differ from what is generally called transfer learning ? I don't think it does differ. (Transfer learning is not a term I'd previously come across). -- Philip Hunt, [EMAIL PROTECTED] Please avoid sending me Word or PowerPoint attachments. See http://www.gnu.org/philosophy/no-word-attachments.html --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244id_secret=120640061-aded06 Powered by Listbox: http://www.listbox.com
Re: [agi] Mushed Up Decision Processes
Ben, Cycorp participated in the DARPA Transfer Learning project, as a subcontractor. My project role was simply a team member and I did not attend any PI meetings. But I did work on getting a Quake III Arena environment working at Cycorp which was to be a transfer learning testbed. I also enhanced Cycorp's Java application that gathered facts from the web using the Google API. Regarding winning a DARPA contract, I believe that teaming with an established contractor, e.g. SAIC, SRI, is beneficial. Cheers, -Steve Stephen L. Reed Artificial Intelligence Researcher http://texai.org/blog http://texai.org 3008 Oak Crest Ave. Austin, Texas, USA 78704 512.791.7860 From: Ben Goertzel [EMAIL PROTECTED] To: agi@v2.listbox.com Sent: Sunday, November 30, 2008 10:17:44 AM Subject: Re: [agi] Mushed Up Decision Processes There was a DARPA program on transfer learning a few years back ... I believe I applied and got rejected (with perfect marks on the technical proposal, as usual ...) ... I never checked to see who got the $$ and what they did with it... ben g On Sun, Nov 30, 2008 at 11:12 AM, Philip Hunt [EMAIL PROTECTED] wrote: 2008/11/30 Ben Goertzel [EMAIL PROTECTED]: Hi, I have proposed a problem domain called function predictor whose purpose is to allow an AI to learn across problem sub-domains, carrying its learning from one domain to another. (See http://www.includipedia.com/wiki/User:Cabalamat/Function_predictor ) I also think it would be useful if there was a regular (maybe annual) competition in the function predictor domain (or some similar domain). A bit like the Loebner Prize, except that it would be more useful to the advancement of AI, since the Loebner prize is silly. -- Philip Hunt, [EMAIL PROTECTED] How does that differ from what is generally called transfer learning ? I don't think it does differ. (Transfer learning is not a term I'd previously come across). -- Philip Hunt, [EMAIL PROTECTED] Please avoid sending me Word or PowerPoint attachments. See http://www.gnu.org/philosophy/no-word-attachments.html --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?; Powered by Listbox: http://www.listbox.com -- Ben Goertzel, PhD CEO, Novamente LLC and Biomind LLC Director of Research, SIAI [EMAIL PROTECTED] I intend to live forever, or die trying. -- Groucho Marx --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?; Powered by Listbox: http://www.listbox.com --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244id_secret=120640061-aded06 Powered by Listbox: http://www.listbox.com
Re: [agi] Mushed Up Decision Processes
Matt Taylor was also an intern at Cycorp where was on Cycorp's Transfer Learning team with me. -Steve Stephen L. Reed Artificial Intelligence Researcher http://texai.org/blog http://texai.org 3008 Oak Crest Ave. Austin, Texas, USA 78704 512.791.7860 From: Pei Wang [EMAIL PROTECTED] To: agi@v2.listbox.com Sent: Sunday, November 30, 2008 10:48:59 AM Subject: Re: [agi] Mushed Up Decision Processes On Sun, Nov 30, 2008 at 11:17 AM, Ben Goertzel [EMAIL PROTECTED] wrote: There was a DARPA program on transfer learning a few years back ... I believe I applied and got rejected (with perfect marks on the technical proposal, as usual ...) ... I never checked to see who got the $$ and what they did with it... See http://www.cs.utexas.edu/~mtaylor/Publications/AGI08-taylor.pdf Pei ben g On Sun, Nov 30, 2008 at 11:12 AM, Philip Hunt [EMAIL PROTECTED] wrote: 2008/11/30 Ben Goertzel [EMAIL PROTECTED]: Hi, I have proposed a problem domain called function predictor whose purpose is to allow an AI to learn across problem sub-domains, carrying its learning from one domain to another. (See http://www.includipedia.com/wiki/User:Cabalamat/Function_predictor ) I also think it would be useful if there was a regular (maybe annual) competition in the function predictor domain (or some similar domain). A bit like the Loebner Prize, except that it would be more useful to the advancement of AI, since the Loebner prize is silly. -- Philip Hunt, [EMAIL PROTECTED] How does that differ from what is generally called transfer learning ? I don't think it does differ. (Transfer learning is not a term I'd previously come across). -- Philip Hunt, [EMAIL PROTECTED] Please avoid sending me Word or PowerPoint attachments. See http://www.gnu.org/philosophy/no-word-attachments.html --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?; Powered by Listbox: http://www.listbox.com -- Ben Goertzel, PhD CEO, Novamente LLC and Biomind LLC Director of Research, SIAI [EMAIL PROTECTED] I intend to live forever, or die trying. -- Groucho Marx --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?; Powered by Listbox: http://www.listbox.com --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?; Powered by Listbox: http://www.listbox.com --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244id_secret=120640061-aded06 Powered by Listbox: http://www.listbox.com
Re: [agi] Mushed Up Decision Processes
Regarding winning a DARPA contract, I believe that teaming with an established contractor, e.g. SAIC, SRI, is beneficial. Cheers, -Steve Yeah, I've tried that approach too ... As it happens, I've had significant more success getting funding from various other government agencies ... but DARPA has been the *least* favorable toward my work of any of them I've tried to deal with It seems that, in the 5 years I've been applying for such grants, DARPA hasn't happened to have a program manager whose particular taste in AI is compatible with mine... -- Ben G --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244id_secret=120640061-aded06 Powered by Listbox: http://www.listbox.com
Re: [agi] Mushed Up Decision Processes
Stephen, Does that mean what you did at Cycorp on transfer learning is similar to what Taylor presented to AGI-08? Pei On Sun, Nov 30, 2008 at 1:01 PM, Stephen Reed [EMAIL PROTECTED] wrote: Matt Taylor was also an intern at Cycorp where was on Cycorp's Transfer Learning team with me. -Steve Stephen L. Reed Artificial Intelligence Researcher http://texai.org/blog http://texai.org 3008 Oak Crest Ave. Austin, Texas, USA 78704 512.791.7860 From: Pei Wang [EMAIL PROTECTED] To: agi@v2.listbox.com Sent: Sunday, November 30, 2008 10:48:59 AM Subject: Re: [agi] Mushed Up Decision Processes On Sun, Nov 30, 2008 at 11:17 AM, Ben Goertzel [EMAIL PROTECTED] wrote: There was a DARPA program on transfer learning a few years back ... I believe I applied and got rejected (with perfect marks on the technical proposal, as usual ...) ... I never checked to see who got the $$ and what they did with it... See http://www.cs.utexas.edu/~mtaylor/Publications/AGI08-taylor.pdf Pei ben g On Sun, Nov 30, 2008 at 11:12 AM, Philip Hunt [EMAIL PROTECTED] wrote: 2008/11/30 Ben Goertzel [EMAIL PROTECTED]: Hi, I have proposed a problem domain called function predictor whose purpose is to allow an AI to learn across problem sub-domains, carrying its learning from one domain to another. (See http://www.includipedia.com/wiki/User:Cabalamat/Function_predictor ) I also think it would be useful if there was a regular (maybe annual) competition in the function predictor domain (or some similar domain). A bit like the Loebner Prize, except that it would be more useful to the advancement of AI, since the Loebner prize is silly. -- Philip Hunt, [EMAIL PROTECTED] How does that differ from what is generally called transfer learning ? I don't think it does differ. (Transfer learning is not a term I'd previously come across). -- Philip Hunt, [EMAIL PROTECTED] Please avoid sending me Word or PowerPoint attachments. See http://www.gnu.org/philosophy/no-word-attachments.html --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?; Powered by Listbox: http://www.listbox.com -- Ben Goertzel, PhD CEO, Novamente LLC and Biomind LLC Director of Research, SIAI [EMAIL PROTECTED] I intend to live forever, or die trying. -- Groucho Marx --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?; Powered by Listbox: http://www.listbox.com --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?; Powered by Listbox: http://www.listbox.com agi | Archives | Modify Your Subscription --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244id_secret=120640061-aded06 Powered by Listbox: http://www.listbox.com
Re: [agi] Mushed Up Decision Processes
Pei, Matt Taylor's work at Cycorp was not closely related to his published work at AGI-08. Matt contributed to a variety of other Transfer Learning tasks, and I cannot recall exactly what those were. -Steve Stephen L. Reed Artificial Intelligence Researcher http://texai.org/blog http://texai.org 3008 Oak Crest Ave. Austin, Texas, USA 78704 512.791.7860 From: Pei Wang [EMAIL PROTECTED] To: agi@v2.listbox.com Sent: Sunday, November 30, 2008 12:16:41 PM Subject: Re: [agi] Mushed Up Decision Processes Stephen, Does that mean what you did at Cycorp on transfer learning is similar to what Taylor presented to AGI-08? Pei On Sun, Nov 30, 2008 at 1:01 PM, Stephen Reed [EMAIL PROTECTED] wrote: Matt Taylor was also an intern at Cycorp where was on Cycorp's Transfer Learning team with me. -Steve Stephen L. Reed Artificial Intelligence Researcher http://texai.org/blog http://texai.org 3008 Oak Crest Ave. Austin, Texas, USA 78704 512.791.7860 From: Pei Wang [EMAIL PROTECTED] To: agi@v2.listbox.com Sent: Sunday, November 30, 2008 10:48:59 AM Subject: Re: [agi] Mushed Up Decision Processes On Sun, Nov 30, 2008 at 11:17 AM, Ben Goertzel [EMAIL PROTECTED] wrote: There was a DARPA program on transfer learning a few years back ... I believe I applied and got rejected (with perfect marks on the technical proposal, as usual ...) ... I never checked to see who got the $$ and what they did with it... See http://www.cs.utexas.edu/~mtaylor/Publications/AGI08-taylor.pdf Pei ben g On Sun, Nov 30, 2008 at 11:12 AM, Philip Hunt [EMAIL PROTECTED] wrote: 2008/11/30 Ben Goertzel [EMAIL PROTECTED]: Hi, I have proposed a problem domain called function predictor whose purpose is to allow an AI to learn across problem sub-domains, carrying its learning from one domain to another. (See http://www.includipedia.com/wiki/User:Cabalamat/Function_predictor ) I also think it would be useful if there was a regular (maybe annual) competition in the function predictor domain (or some similar domain). A bit like the Loebner Prize, except that it would be more useful to the advancement of AI, since the Loebner prize is silly. -- Philip Hunt, [EMAIL PROTECTED] How does that differ from what is generally called transfer learning ? I don't think it does differ. (Transfer learning is not a term I'd previously come across). -- Philip Hunt, [EMAIL PROTECTED] Please avoid sending me Word or PowerPoint attachments. See http://www.gnu.org/philosophy/no-word-attachments.html --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?; Powered by Listbox: http://www.listbox.com -- Ben Goertzel, PhD CEO, Novamente LLC and Biomind LLC Director of Research, SIAI [EMAIL PROTECTED] I intend to live forever, or die trying. -- Groucho Marx --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?; Powered by Listbox: http://www.listbox.com --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?; Powered by Listbox: http://www.listbox.com agi | Archives | Modify Your Subscription --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?; Powered by Listbox: http://www.listbox.com --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244id_secret=120640061-aded06 Powered by Listbox: http://www.listbox.com
Re: [agi] Mushed Up Decision Processes
Ed, I think that we must rely on large collections of relatively simple patterns that are somehow capable of being mixed and used in interactions with the others. These interacting patterns (to use your term) would have extensive variations to make them flexible and useful with other patterns. When we learn that national housing prices did not provide us with the kind of detail that we needed we go and figure other ways to find data that showed some of the variations that would have helped us to prepare better for a situation like the one we are currently in. I was thinking of that exact example when I wrote about mushy decision making, because the national average price would be more mushy than the regional prices, or a multiple price level index. The mush index of an index does not mean that the index is garbage, but since something like this is derived from finer grained statistics, it really exemplifies the problem. My idea is that an agi program would have to go further than data mining. It would have to be able to shape its own use of statistics in order to establish validity for itself. I really feel that there is something really important about the classifiers of statistical methods that I just haven't grasped yet. My example for this this comes from statistics that are similar but just different enough so that they don't mesh quite right. Like two different marketing surveys that provide similar information which is so close that a marketer can draw conclusions from their combination but which aren't actually close enough to justify this process. Like asking different representative groups if they are planning to buy a television in one survey, and asking how much they think they will spend on appliances during the next two years. The two surveys are so close that you know the results can be combined, but they are so different that it is almost impossible to justify the combination in any reasonable way. If I could only figure this one out I think the other problems I am interested in would start to solve themselves. Jim Bromer On Sat, Nov 29, 2008 at 11:40 AM, Ed Porter [EMAIL PROTECTED] wrote: Jim My understanding is that a Novamente-like system would have a process of natural selection that tends to favor the retention and use of patterns (perceptive, cognative, behaviors) prove themselves useful in achieving goals in the word in which it is embodied. It seems to me t such a process of natural selection would tend to naturally put some sort of limit on how out-of-touch many of an AGI's patterns would be, at least with regard to patterns about things for which the AGI has had considerable experience from the world in which it is embodied. However, we humans often get pretty out of touch with real world probabilities, as the recent bubble in housing prices, and the commonly said, although historically inaccurate, statement of several years ago --- that housing prices never go down on a national --- shows. It would be helpful to make AGI's be a little more accurate in their evaluation of the evidence for many of their assumptions --- and what that evidence really says --- than we humans are. Ed Porter -Original Message- From: Jim Bromer [mailto:[EMAIL PROTECTED] Sent: Saturday, November 29, 2008 10:49 AM To: agi@v2.listbox.com Subject: [agi] Mushed Up Decision Processes One of the problems that comes with the casual use of analytical methods is that the user becomes inured to their habitual misuse. When a casual familiarity is combined with a habitual ignorance of the consequences of a misuse the user can become over-confident or unwisely dismissive of criticism regardless of how on the mark it might be. The most proper use of statistical and probabilistic methods is to base results on a strong association with the data that they were derived from. The problem is that the AI community cannot afford this strong a connection to original source because they are trying to emulate the mind in some way and it is not reasonable to assume that the mind is capable of storing all data that it has used to derive insight. This is a problem any AI method has to deal with, it is not just a probability thing. What is wrong with the AI-probability group mind-set is that very few of its proponents ever consider the problem of statistical ambiguity and its obvious consequences. All AI programmers have to consider the problem. Most theories about the mind posit the use of similar experiences to build up theories about the world (or to derive methods to deal effectively with the world). So even though the methods to deal with the data environment are detached from the original sources of those methods, they can still be reconnected by the examination of similar experiences that may subsequently occur. But still it is important to be able to recognize the significance and necessity of doing this from time to time. It is important to
Re: [agi] Mushed Up Decision Processes
On Nov 30, 2008, at 7:31 AM, Philip Hunt wrote: 2008/11/30 Ben Goertzel [EMAIL PROTECTED]: In general, the standard AI methods can't handle pattern recognition problems requiring finding complex interdependencies among multiple variables that are obscured among scads of other variables The human mind seems to do this via building up intuition via drawing analogies among multiple problems it confronts during its history. Yes, so that people learn one problem, then it helps them to learn other similar ones. Is there any AI software that does this? I'm not aware of any. To do this as a practical matter, you need to address *at least* two well-known hard-but-important unsolved algorithm problems in completely different areas of theoretical computer science that have nothing to do with AI per se. That is no small hurdle, even if you are a bloody genius. That said, I doubt most AI researchers could even tell you what those two big problems are which is, obliquely, the other part of the problem. I have proposed a problem domain called function predictor whose purpose is to allow an AI to learn across problem sub-domains, carrying its learning from one domain to another. (See http://www.includipedia.com/wiki/User:Cabalamat/Function_predictor ) In Feder/Merhav/Gutman's 1995 Reflections on... followup to their 1992 paper on universal sequence prediction, they make the observation, which can be found at the following link, that it is probably useful to introduce the concept of prediction error complexity as an important metric which is similar to what you are talking about in the theoretical abstract: http://www.itsoc.org/review/meir/node5.html Our understanding of this area is better in 2008 than it was in 1995, but this is one of the earliest serious references to the idea in a theoretical way. Somewhat obscure and primitive by current standards, but influential in the AIXI and related flavors of AI theory based on computational information theory. Or at least, I found it very interesting and useful a decade ago. Cheers, J. Andrew Rogers --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244id_secret=120640061-aded06 Powered by Listbox: http://www.listbox.com
Re: [agi] Mushed Up Decision Processes
Jim, There is a large body of literature on avoiding overfitting, ie, finding patterns that work for more then just the data at hand. Of course, the ultimate conclusion is that you can never be 100% sure; but some interesting safeguards have been cooked up anyway, which help in practice. My point is, the following paragraph is unfounded: This is a problem any AI method has to deal with, it is not just a probability thing. What is wrong with the AI-probability group mind-set is that very few of its proponents ever consider the problem of statistical ambiguity and its obvious consequences. The AI-probability group definitely considers such problems. --Abram On Sat, Nov 29, 2008 at 10:48 AM, Jim Bromer [EMAIL PROTECTED] wrote: One of the problems that comes with the casual use of analytical methods is that the user becomes inured to their habitual misuse. When a casual familiarity is combined with a habitual ignorance of the consequences of a misuse the user can become over-confident or unwisely dismissive of criticism regardless of how on the mark it might be. The most proper use of statistical and probabilistic methods is to base results on a strong association with the data that they were derived from. The problem is that the AI community cannot afford this strong a connection to original source because they are trying to emulate the mind in some way and it is not reasonable to assume that the mind is capable of storing all data that it has used to derive insight. This is a problem any AI method has to deal with, it is not just a probability thing. What is wrong with the AI-probability group mind-set is that very few of its proponents ever consider the problem of statistical ambiguity and its obvious consequences. All AI programmers have to consider the problem. Most theories about the mind posit the use of similar experiences to build up theories about the world (or to derive methods to deal effectively with the world). So even though the methods to deal with the data environment are detached from the original sources of those methods, they can still be reconnected by the examination of similar experiences that may subsequently occur. But still it is important to be able to recognize the significance and necessity of doing this from time to time. It is important to be able to reevaluate parts of your theories about things. We are not just making little modifications from our internal theories about things when we react to ongoing events, we must be making some sort of reevaluation of our insights about the kind of thing that we are dealing with as well. I realize now that most people in these groups probably do not understand where I am coming from because their idea of AI programming is based on a model of programming that is flat. You have the program at one level and the possible reactions to the data that is input as the values of the program variables are carefully constrained by that level. You can imagine a more complex model of programming by appreciating the possibility that the program can react to IO data by rearranging subprograms to make new kinds of programs. Although a subtle argument can be made that any program that conditionally reacts to input data is rearranging the execution of its subprograms, the explicit recognition by the programmer that this is useful tool in advanced programming is probably highly correlated with its more effective use. (I mean of course it is highly correlated with its effective use!) I believe that casually constructed learning methods (and decision processes) can lead to even more uncontrollable results when used with this self-programming aspect of advanced AI programs. The consequences then of failing to recognize that mushed up decision processes that are never compared against the data (or kinds of situations) that they were derived from will be the inevitable emergence of inherently illogical decision processes that will mush up an AI system long before it gets any traction. Jim Bromer --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?; Powered by Listbox: http://www.listbox.com --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244id_secret=120640061-aded06 Powered by Listbox: http://www.listbox.com
Re: [agi] Mushed Up Decision Processes
Jim, YES - and I think I have another piece of your puzzle to consider... A longtime friend of mine, Dave, went on to become a PhD psychologist, who subsequently took me on as a sort of project - to figure out why most people who met me then either greatly valued my friendship, or quite the opposite, would probably kill me if they had the safe opportunity. After much discussion, interviewing people in both camps, etc., he came up with what appears to be a key to decision making in general... It appears that people pigeonhole other people, concepts, situations, etc., into a very finite number of pigeonholes - probably just tens of pigeonholes for other people. Along with the pigeonhole, they keep amendments, like Steve is like Joe, but with Then, there is the pigeonhole labeled other that all the mavericks are thrown into. Not being at all like anyone else that most people have ever met, I was invariably filed into the other pigeonhole, along with Einstein, Ted Bundy, Jack the Ripper, Stephen Hawking, etc. People are safe to the extent that they are predictable, and people in the other pigeonhole got that way because they appear to NOT be predictable, e.g. because of their worldview, etc. Now, does the potential value of the alternative worldview outweigh the potential danger of perceived unpredictability? The answer to this question apparently drove my own personal classification in other people. Dave's goal was to devise a way to stop making enemies, but unfortunately, this model of how people got that way suggested no potential solution. People who keep themselves safe from others having radically different worldviews are truly in a mental prison of their own making, and there is no way that someone whom they distrust could ever release them from that prison. I suspect that recognition, decision making, and all sorts of intelligent processes may be proceeding in much the same way. There may be no grandmother neuron/pidgeonhole, but rather a kindly old person with an amendment that is related. If on the other hand your other grandmother flogged you as a child, the filing might be quite different. Any thoughts? Steve Richfield On 11/29/08, Jim Bromer [EMAIL PROTECTED] wrote: One of the problems that comes with the casual use of analytical methods is that the user becomes inured to their habitual misuse. When a casual familiarity is combined with a habitual ignorance of the consequences of a misuse the user can become over-confident or unwisely dismissive of criticism regardless of how on the mark it might be. The most proper use of statistical and probabilistic methods is to base results on a strong association with the data that they were derived from. The problem is that the AI community cannot afford this strong a connection to original source because they are trying to emulate the mind in some way and it is not reasonable to assume that the mind is capable of storing all data that it has used to derive insight. This is a problem any AI method has to deal with, it is not just a probability thing. What is wrong with the AI-probability group mind-set is that very few of its proponents ever consider the problem of statistical ambiguity and its obvious consequences. All AI programmers have to consider the problem. Most theories about the mind posit the use of similar experiences to build up theories about the world (or to derive methods to deal effectively with the world). So even though the methods to deal with the data environment are detached from the original sources of those methods, they can still be reconnected by the examination of similar experiences that may subsequently occur. But still it is important to be able to recognize the significance and necessity of doing this from time to time. It is important to be able to reevaluate parts of your theories about things. We are not just making little modifications from our internal theories about things when we react to ongoing events, we must be making some sort of reevaluation of our insights about the kind of thing that we are dealing with as well. I realize now that most people in these groups probably do not understand where I am coming from because their idea of AI programming is based on a model of programming that is flat. You have the program at one level and the possible reactions to the data that is input as the values of the program variables are carefully constrained by that level. You can imagine a more complex model of programming by appreciating the possibility that the program can react to IO data by rearranging subprograms to make new kinds of programs. Although a subtle argument can be made that any program that conditionally reacts to input data is rearranging the execution of its subprograms, the explicit recognition by the programmer that this is useful tool in advanced programming is probably highly correlated with its more effective use.
Re: [agi] Mushed Up Decision Processes
Hi. I will just make a quick response to this message and then I want to think about the other messages before I reply. A few weeks ago I decided that I would write a criticism of ai-probability to post to this group. I wasn't able remember all of my criticisms so I decided to post a few preliminary sketches to another group. I wasn't too concerned about how they responded, and in fact I thought they would just ignore me. The first response I got was from an irate guy who was quite unpleasant and then finished by declaring that I slandered the entire ai-probability community! He had some reasonable criticisms about this but I considered the issue tangential to the central issue I wanted to discuss. I would have responded to his more reasonable criticisms if they hadn't been embedded in his enraged rant. I wondered why anyone would deface the expression of his own thoughts with an emotional and hostile message, so I wanted to try the same message on this group to see if anyone who was more mature would focus on this same issue. Abram made a measured response but his focus was on the over-generalization. As I said, this was just a preliminary sketch of a message that I intended to post to this group after I had worked on it. Your point is taken. Norvig seems to say that overfitting is a general problem. The method given to study the problem is probabilistic but it is based on the premise that the original data is substantially intact. But Norvig goes on to mention that with pruning noise can be tolerated. If you read my message again you may see that my central issue was not really centered on the issue of whether anyone in the ai-probability community was aware of the nature of the science of statistics but whether or not probability can be used as the fundamental basis to create agi given the complexities of the problem. So while your example of overfitting certainly does deflate my statements that no one in the ai-probability community gets this stuff, it does not actually address the central issue that I was thinking of. I am not sure if Norvig's application of a probabilistic method to detect overfitting is truly directed toward the agi community. In other words: Has anyone in this grouped tested the utility and clarity of the decision making of a fully automated system to detect overfitting in a range of complex IO data fields that one might expect to encounter in AGI? Jim Bromer On Sat, Nov 29, 2008 at 11:32 AM, Abram Demski [EMAIL PROTECTED] wrote: Jim, There is a large body of literature on avoiding overfitting, ie, finding patterns that work for more then just the data at hand. Of course, the ultimate conclusion is that you can never be 100% sure; but some interesting safeguards have been cooked up anyway, which help in practice. My point is, the following paragraph is unfounded: This is a problem any AI method has to deal with, it is not just a probability thing. What is wrong with the AI-probability group mind-set is that very few of its proponents ever consider the problem of statistical ambiguity and its obvious consequences. The AI-probability group definitely considers such problems. --Abram On Sat, Nov 29, 2008 at 10:48 AM, Jim Bromer [EMAIL PROTECTED] wrote: One of the problems that comes with the casual use of analytical methods is that the user becomes inured to their habitual misuse. When a casual familiarity is combined with a habitual ignorance of the consequences of a misuse the user can become over-confident or unwisely dismissive of criticism regardless of how on the mark it might be. The most proper use of statistical and probabilistic methods is to base results on a strong association with the data that they were derived from. The problem is that the AI community cannot afford this strong a connection to original source because they are trying to emulate the mind in some way and it is not reasonable to assume that the mind is capable of storing all data that it has used to derive insight. This is a problem any AI method has to deal with, it is not just a probability thing. What is wrong with the AI-probability group mind-set is that very few of its proponents ever consider the problem of statistical ambiguity and its obvious consequences. All AI programmers have to consider the problem. Most theories about the mind posit the use of similar experiences to build up theories about the world (or to derive methods to deal effectively with the world). So even though the methods to deal with the data environment are detached from the original sources of those methods, they can still be reconnected by the examination of similar experiences that may subsequently occur. But still it is important to be able to recognize the significance and necessity of doing this from time to time. It is important to be able to reevaluate parts of your theories about things. We are not just making little modifications from our
Re: [agi] Mushed Up Decision Processes
Well, if you're willing to take the step of asking questions about the world that are framed in terms of probabilities and probability distributions ... then modern probability and statistics tell you a lot about overfitting and how to avoid it... OTOH if, like Pei Wang, you think it's misguided to ask questions posed in a probabilistic framework, then that theory will not be directly relevant to you... To me the big weaknesses of modern probability theory lie in **hypothesis generation** and **inference**. Testing a hypothesis against data, to see if it's overfit to that data, is handled well by crossvalidation and related methods. But the problem of: given a number of hypotheses with support from a dataset, generating other interesting hypotheses that will also have support from the dataset ... that is where traditional probabilistic methods (though not IMO the foundational ideas of probability) fall short, providing only unscalable or oversimplified solutions... -- Ben G On Sat, Nov 29, 2008 at 1:08 PM, Jim Bromer [EMAIL PROTECTED] wrote: Hi. I will just make a quick response to this message and then I want to think about the other messages before I reply. A few weeks ago I decided that I would write a criticism of ai-probability to post to this group. I wasn't able remember all of my criticisms so I decided to post a few preliminary sketches to another group. I wasn't too concerned about how they responded, and in fact I thought they would just ignore me. The first response I got was from an irate guy who was quite unpleasant and then finished by declaring that I slandered the entire ai-probability community! He had some reasonable criticisms about this but I considered the issue tangential to the central issue I wanted to discuss. I would have responded to his more reasonable criticisms if they hadn't been embedded in his enraged rant. I wondered why anyone would deface the expression of his own thoughts with an emotional and hostile message, so I wanted to try the same message on this group to see if anyone who was more mature would focus on this same issue. Abram made a measured response but his focus was on the over-generalization. As I said, this was just a preliminary sketch of a message that I intended to post to this group after I had worked on it. Your point is taken. Norvig seems to say that overfitting is a general problem. The method given to study the problem is probabilistic but it is based on the premise that the original data is substantially intact. But Norvig goes on to mention that with pruning noise can be tolerated. If you read my message again you may see that my central issue was not really centered on the issue of whether anyone in the ai-probability community was aware of the nature of the science of statistics but whether or not probability can be used as the fundamental basis to create agi given the complexities of the problem. So while your example of overfitting certainly does deflate my statements that no one in the ai-probability community gets this stuff, it does not actually address the central issue that I was thinking of. I am not sure if Norvig's application of a probabilistic method to detect overfitting is truly directed toward the agi community. In other words: Has anyone in this grouped tested the utility and clarity of the decision making of a fully automated system to detect overfitting in a range of complex IO data fields that one might expect to encounter in AGI? Jim Bromer On Sat, Nov 29, 2008 at 11:32 AM, Abram Demski [EMAIL PROTECTED] wrote: Jim, There is a large body of literature on avoiding overfitting, ie, finding patterns that work for more then just the data at hand. Of course, the ultimate conclusion is that you can never be 100% sure; but some interesting safeguards have been cooked up anyway, which help in practice. My point is, the following paragraph is unfounded: This is a problem any AI method has to deal with, it is not just a probability thing. What is wrong with the AI-probability group mind-set is that very few of its proponents ever consider the problem of statistical ambiguity and its obvious consequences. The AI-probability group definitely considers such problems. --Abram On Sat, Nov 29, 2008 at 10:48 AM, Jim Bromer [EMAIL PROTECTED] wrote: One of the problems that comes with the casual use of analytical methods is that the user becomes inured to their habitual misuse. When a casual familiarity is combined with a habitual ignorance of the consequences of a misuse the user can become over-confident or unwisely dismissive of criticism regardless of how on the mark it might be. The most proper use of statistical and probabilistic methods is to base results on a strong association with the data that they were derived from. The problem is that the AI community cannot afford this strong a connection to original source because they are
Re: [agi] Mushed Up Decision Processes
--- On Sat, 11/29/08, Jim Bromer [EMAIL PROTECTED] wrote: I am not sure if Norvig's application of a probabilistic method to detect overfitting is truly directed toward the agi community. In other words: Has anyone in this grouped tested the utility and clarity of the decision making of a fully automated system to detect overfitting in a range of complex IO data fields that one might expect to encounter in AGI? The general problem of detecting overfitting is not computable. The principle according to Occam's Razor, formalized and proven by Hutter's AIXI model, is to choose the shortest program (simplest hypothesis) that generates the data. Overfitting is the case of choosing a program that is too large. -- Matt Mahoney, [EMAIL PROTECTED] --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244id_secret=120640061-aded06 Powered by Listbox: http://www.listbox.com
Re: [agi] Mushed Up Decision Processes
A response to: I wondered why anyone would deface the expression of his own thoughts with an emotional and hostile message, My theory is that thoughts are generated internally and forced into words via a babble generator. Then the thoughts are filtered through a screen to remove any that don't match ones intent, that don't make sense, etc. The value assigned to each expression is initially dependent on how well it expresses one's emotional tenor. Therefore I would guess that all of the verbalizations that the individual generated which passed the first screen were hostile in nature. From the remaining sample he filtered those which didn't generate sensible-to-him scenarios when fed back into his world model. This left him with a much reduced selection of phrases to choose from when composing his response. In my model this happens a phrase at a time rather than a sentence at a time. And there is also a probabilistic element where each word has a certain probability of being followed by divers other words. I often don't want to express the most likely probability, as by choosing a less frequently chosen alternative I (believe I) create the impression a more studied, i.e. thoughtful, response. But if one wishes to convey a more dynamic style then one would choose a more likely follower. Note that in this scenario phrases are generated both randomly and in parallel. Then they are selected for fitness for expression by passing through various filter. Reasonable? Jim Bromer wrote: Hi. I will just make a quick response to this message and then I want to think about the other messages before I reply. A few weeks ago I decided that I would write a criticism of ai-probability to post to this group. I wasn't able remember all of my criticisms so I decided to post a few preliminary sketches to another group. I wasn't too concerned about how they responded, and in fact I thought they would just ignore me. The first response I got was from an irate guy who was quite unpleasant and then finished by declaring that I slandered the entire ai-probability community! He had some reasonable criticisms about this but I considered the issue tangential to the central issue I wanted to discuss. I would have responded to his more reasonable criticisms if they hadn't been embedded in his enraged rant. I wondered why anyone would deface the expression of his own thoughts with an emotional and hostile message, so I wanted to try the same message on this group to see if anyone who was more mature would focus on this same issue. Abram made a measured response but his focus was on the over-generalization. As I said, this was just a preliminary sketch of a message that I intended to post to this group after I had worked on it. Your point is taken. Norvig seems to say that overfitting is a general problem. The method given to study the problem is probabilistic but it is based on the premise that the original data is substantially intact. But Norvig goes on to mention that with pruning noise can be tolerated. If you read my message again you may see that my central issue was not really centered on the issue of whether anyone in the ai-probability community was aware of the nature of the science of statistics but whether or not probability can be used as the fundamental basis to create agi given the complexities of the problem. So while your example of overfitting certainly does deflate my statements that no one in the ai-probability community gets this stuff, it does not actually address the central issue that I was thinking of. I am not sure if Norvig's application of a probabilistic method to detect overfitting is truly directed toward the agi community. In other words: Has anyone in this grouped tested the utility and clarity of the decision making of a fully automated system to detect overfitting in a range of complex IO data fields that one might expect to encounter in AGI? Jim Bromer On Sat, Nov 29, 2008 at 11:32 AM, Abram Demski [EMAIL PROTECTED] wrote: Jim, There is a large body of literature on avoiding overfitting, ie, finding patterns that work for more then just the data at hand. Of course, the ultimate conclusion is that you can never be 100% sure; but some interesting safeguards have been cooked up anyway, which help in practice. My point is, the following paragraph is unfounded: This is a problem any AI method has to deal with, it is not just a probability thing. What is wrong with the AI-probability group mind-set is that very few of its proponents ever consider the problem of statistical ambiguity and its obvious consequences. The AI-probability group definitely considers such problems. --Abram On Sat, Nov 29, 2008 at 10:48 AM, Jim Bromer [EMAIL PROTECTED] wrote: One of the problems that comes with the casual use of analytical methods is that the user becomes inured to their habitual misuse. When a casual familiarity is combined with a
Re: [agi] Mushed Up Decision Processes
In response to my message, where I said, What is wrong with the AI-probability group mind-set is that very few of its proponents ever consider the problem of statistical ambiguity and its obvious consequences. Abram noted, The AI-probability group definitely considers such problems. There is a large body of literature on avoiding overfitting, ie, finding patterns that work for more then just the data at hand. Suppose I responded with a remark like, 6341/6344 wrong Abram... A remark like this would be absurd because it lacks reference, explanation and validity while also presenting a comically false numerical precision for its otherwise inherent meaninglessness. Where does the ratio 6341/6344 come from? I did a search in ListBox of all references to the word overfitting made in 2008 and found that out of 6344 messages only 3 actually involved the discussion of the word before Abram mentioned it today. (I don't know how good ListBox is for this sort of thing). So what is wrong with my conclusion that Abram was 6341/6344 wrong? Lots of things and they can all be described using declarative statements. First of all the idea that the conversations in this newsgroup represent an adequate sampling of all ai-probability enthusiasts is totally ridiculous. Secondly, Abram's mention of overfitting was just one example of how the general ai-probability community is aware of the problem that I mentioned. So while my statistical finding may be tangentially relevant to the discussion, the presumption that it can serve as a numerical evaluation of Abram's 'wrongness' in his response is so absurd that it does not merit serious consideration. My skepticism then concerns the question of just how would a fully automated AGI program that relied fully on probability methods be able to avoid getting sucked into the vortex of such absurd mushy reasoning if it wasn't also able to analyze the declarative inferences of its application of statistical methods? I believe that an AI program that is to be capable of advanced AGI has to be capable of declarative assessment to work with any other mathematical methods of reasoning it is programmed with. The ability to reason about declarative knowledge does not necessarily have to be done in text or something like that. That is not what I mean. What I really mean is that an effective AI program is going to have to be capable of some kind of referential analysis of events in the IO data environment using methods other than probability. But if it is to attain higher intellectual functions it has to be done in a creative and imaginative way. Just as human statisticians have to be able to express and analyze the application of their statistical methods using declarative statements that refer to the data subject fields and the methods used, an AI program that is designed to utilize automated probability reasoning to attain greater general success is going to have to be able to express and analyze its statistical assessments in terms of some kind of declarative methods as well. Jim Bromer --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244id_secret=120640061-aded06 Powered by Listbox: http://www.listbox.com
Re: [agi] Mushed Up Decision Processes
On Sat, Nov 29, 2008 at 1:51 PM, Ben Goertzel [EMAIL PROTECTED] wrote: To me the big weaknesses of modern probability theory lie in **hypothesis generation** and **inference**. Testing a hypothesis against data, to see if it's overfit to that data, is handled well by crossvalidation and related methods. But the problem of: given a number of hypotheses with support from a dataset, generating other interesting hypotheses that will also have support from the dataset ... that is where traditional probabilistic methods (though not IMO the foundational ideas of probability) fall short, providing only unscalable or oversimplified solutions... -- Ben G Could you give me a little more detail about your thoughts on this? Do you think the problem of increasing uncomputableness of complicated complexity is the common thread found in all of the interesting, useful but unscalable methods of AI? Jim Bromer --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244id_secret=120640061-aded06 Powered by Listbox: http://www.listbox.com
Re: [agi] Mushed Up Decision Processes
Whether an AI needs to explicitly manipulate declarative statements is a deep question ... it may be that other dynamics that are in some contexts implicitly equivalent to this sort of manipulation will suffice But anyway, there is no contradiction between manipulating explicit declarative statements and using probability theory. Some of my colleagues and I spent a bunch of time during the last few years figuring out nice ways to combine probability theory and formal logic. In fact there are Progic workshops every year exploring these sorts of themes. So, while the mainstream of probability-focused AI theorists aren't doing hard-core probabilistic logic, some researchers certainly are... I've been displeased with the wimpiness of the progic subfield, and its lack of contribution to areas like inference with nested quantifiers, and intensional inference ... and I've tried to remedy these shortcomings with PLN (Probabilistic Logic Networks) ... So, I think it's correct to criticize the mainstream of probability-focused AI theorists for not doing AGI ;-) ... but I don't think they've overlooking basic issues like overfitting and such ... I think they're just focusing on relatively easy problems where (unlike if you want to do explicitly probability theory based AGI) you don't need to merge probability theory with complex logical constructs... ben On Sat, Nov 29, 2008 at 9:15 PM, Jim Bromer [EMAIL PROTECTED] wrote: In response to my message, where I said, What is wrong with the AI-probability group mind-set is that very few of its proponents ever consider the problem of statistical ambiguity and its obvious consequences. Abram noted, The AI-probability group definitely considers such problems. There is a large body of literature on avoiding overfitting, ie, finding patterns that work for more then just the data at hand. Suppose I responded with a remark like, 6341/6344 wrong Abram... A remark like this would be absurd because it lacks reference, explanation and validity while also presenting a comically false numerical precision for its otherwise inherent meaninglessness. Where does the ratio 6341/6344 come from? I did a search in ListBox of all references to the word overfitting made in 2008 and found that out of 6344 messages only 3 actually involved the discussion of the word before Abram mentioned it today. (I don't know how good ListBox is for this sort of thing). So what is wrong with my conclusion that Abram was 6341/6344 wrong? Lots of things and they can all be described using declarative statements. First of all the idea that the conversations in this newsgroup represent an adequate sampling of all ai-probability enthusiasts is totally ridiculous. Secondly, Abram's mention of overfitting was just one example of how the general ai-probability community is aware of the problem that I mentioned. So while my statistical finding may be tangentially relevant to the discussion, the presumption that it can serve as a numerical evaluation of Abram's 'wrongness' in his response is so absurd that it does not merit serious consideration. My skepticism then concerns the question of just how would a fully automated AGI program that relied fully on probability methods be able to avoid getting sucked into the vortex of such absurd mushy reasoning if it wasn't also able to analyze the declarative inferences of its application of statistical methods? I believe that an AI program that is to be capable of advanced AGI has to be capable of declarative assessment to work with any other mathematical methods of reasoning it is programmed with. The ability to reason about declarative knowledge does not necessarily have to be done in text or something like that. That is not what I mean. What I really mean is that an effective AI program is going to have to be capable of some kind of referential analysis of events in the IO data environment using methods other than probability. But if it is to attain higher intellectual functions it has to be done in a creative and imaginative way. Just as human statisticians have to be able to express and analyze the application of their statistical methods using declarative statements that refer to the data subject fields and the methods used, an AI program that is designed to utilize automated probability reasoning to attain greater general success is going to have to be able to express and analyze its statistical assessments in terms of some kind of declarative methods as well. Jim Bromer --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?; Powered by Listbox: http://www.listbox.com -- Ben Goertzel, PhD CEO, Novamente LLC and Biomind LLC Director of Research, SIAI [EMAIL PROTECTED] I intend to live forever, or die trying. --
Re: [agi] Mushed Up Decision Processes
Could you give me a little more detail about your thoughts on this? Do you think the problem of increasing uncomputableness of complicated complexity is the common thread found in all of the interesting, useful but unscalable methods of AI? Jim Bromer Well, I think that dealing with combinatorial explosions is, in general, the great unsolved problem of AI. I think the opencog prime design can solve it, but this isn't proved yet... Even relatively unambitious AI methods tend to get dumbed down further when you try to scale them up, due to combinatorial explosion issues. For instance, Bayes nets aren't that clever to begin with ... they don't do that much ... but to make them scalable, one has to make them even more limited and basically ignore combinational causes and just look at causes between one isolated event-class and another... And of course, all theorem provers are unscalable due to having no scalable methods of inference tree pruning... Evolutionary methods can't handle complex fitness functions because they'd require overly large population sizes... In general, the standard AI methods can't handle pattern recognition problems requiring finding complex interdependencies among multiple variables that are obscured among scads of other variables The human mind seems to do this via building up intuition via drawing analogies among multiple problems it confronts during its history. Also of course the human mind builds internal simulations of the world, and probes these simulations and draws analogies from problems it solved in its inner sim world, to problems it encounters in the outer world... etc. etc. etc. ben --- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244id_secret=120640061-aded06 Powered by Listbox: http://www.listbox.com
Re: [agi] Mushed Up Decision Processes
On Sat, Nov 29, 2008 at 11:53 AM, Steve Richfield [EMAIL PROTECTED] wrote: Jim, YES - and I think I have another piece of your puzzle to consider... A longtime friend of mine, Dave, went on to become a PhD psychologist, who subsequently took me on as a sort of project - to figure out why most people who met me then either greatly valued my friendship, or quite the opposite, would probably kill me if they had the safe opportunity. After much discussion, interviewing people in both camps, etc., he came up with what appears to be a key to decision making in general... It appears that people pigeonhole other people, concepts, situations, etc., into a very finite number of pigeonholes - probably just tens of pigeonholes for other people. Steve: I found that I used a similar method of categorizing people who I talked to on these newsgroups. I wouldn't call it pigeonholing though. (Actually, I wouldn't call anything pigeonholing, but that is just me.) I would rely on a handful of generalizations that I thought were applicable to different people who tended to exhibit some common characteristics. However, when I discovered that an individual who I thought I understood had another facet to his personality or thoughts that I hadn't seen before I often found that I had to apply another categorical generality to my impression of him. I soon built up generalization categories based on different experiences with different kinds of people, and I eventually realized that although I often saw similar kinds of behaviors in different people, each person seemed to be comprised of different sets (or different strengths) of the various component characteristics that I derived to recall my experiences with people in these groups. So I came to similar conclusions that you and your friend came to. An interesting thing about talking to reactive people in these discussion groups. I found that by eliminating more and more affect from my comments, by refraining from personal comments, innuendos or making meta-discussion analyses and by increasingly emphasizing objectivity in my comments I could substantially reduce any hostility directed at me. My problem is that I do not want to remove all affect from my conversation just to placate some unpleasant person. But I guess I should start using that technique again when necessary. Jim Bromer On Sat, Nov 29, 2008 at 11:53 AM, Steve Richfield [EMAIL PROTECTED] wrote: Jim, YES - and I think I have another piece of your puzzle to consider... A longtime friend of mine, Dave, went on to become a PhD psychologist, who subsequently took me on as a sort of project - to figure out why most people who met me then either greatly valued my friendship, or quite the opposite, would probably kill me if they had the safe opportunity. After much discussion, interviewing people in both camps, etc., he came up with what appears to be a key to decision making in general... It appears that people pigeonhole other people, concepts, situations, etc., into a very finite number of pigeonholes - probably just tens of pigeonholes for other people. Along with the pigeonhole, they keep amendments, like Steve is like Joe, but with Then, there is the pigeonhole labeled other that all the mavericks are thrown into. Not being at all like anyone else that most people have ever met, I was invariably filed into the other pigeonhole, along with Einstein, Ted Bundy, Jack the Ripper, Stephen Hawking, etc. People are safe to the extent that they are predictable, and people in the other pigeonhole got that way because they appear to NOT be predictable, e.g. because of their worldview, etc. Now, does the potential value of the alternative worldview outweigh the potential danger of perceived unpredictability? The answer to this question apparently drove my own personal classification in other people. Dave's goal was to devise a way to stop making enemies, but unfortunately, this model of how people got that way suggested no potential solution. People who keep themselves safe from others having radically different worldviews are truly in a mental prison of their own making, and there is no way that someone whom they distrust could ever release them from that prison. I suspect that recognition, decision making, and all sorts of intelligent processes may be proceeding in much the same way. There may be no grandmother neuron/pidgeonhole, but rather a kindly old person with an amendment that is related. If on the other hand your other grandmother flogged you as a child, the filing might be quite different. Any thoughts? Steve Richfield On 11/29/08, Jim Bromer [EMAIL PROTECTED] wrote: One of the problems that comes with the casual use of analytical methods is that the user becomes inured to their habitual misuse. When a casual familiarity is combined with a habitual ignorance of the consequences of a misuse the user can become