Re: Mark Launches Singularity :-) WAS Re: [agi] Microsoft Launches Singularity

2008-03-26 Thread Chris Petersen
Mentifex called; it wants its ASCII diagrams back.

-Chris

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


Re: [agi] Instead of an AGI Textbook

2008-03-26 Thread Aki Iskandar
Well ... I can take a shot at putting a diagram together.  Making Mind Maps
is one way I learn any kind of material I want.

If the topics in the list(s) are descriptive enough, I can take a shot at
putting such a diagram together.
It'd be less work to correct it than to make one, right?

Hey - whatever helps.  For me, it's a win-win.  It would help me, and it
would help accomplish what you guys are trying to do.

Let me know,
~Aki

On Tue, Mar 25, 2008 at 10:40 PM, Ben Goertzel [EMAIL PROTECTED] wrote:

 This kind of diagram would certainly be meaningful, but, it would be a
 lot of work to put together, even more so than a traditional TOC ...

 On Tue, Mar 25, 2008 at 11:02 PM, Aki Iskandar [EMAIL PROTECTED] wrote:
  Hi Pei -
 
  What about having a tree like diagram that branches out into either:
  - the different paths / approaches to AGI (for instance: NARS,
 Novamente,
  and Richard's, etc.), with suggested readings at those leaves
   - area of study, with suggested readings at those leaves
 
  Or possibly, a Mind Map diagram that shows AGI in the middle, with the
  approaches stemming from it, and then either sub fields, or a reading
 list
  and / or collection of links (though the links may become outdated,
 dead).
 
  Point is, would a diagram help map the field - which caters to the
  differing approaches, and which helps those wanting to chart a course to
  their own learning/study ?
 
  Thanks,
  ~Aki
 
 
 
 
   On Tue, Mar 25, 2008 at 9:22 PM, Pei Wang [EMAIL PROTECTED]
 wrote:
   Ben,
  
   It is a good start!
  
   Of course everyone else will disagree --- like what Richard did and
   I'm going to do. ;-)
  
   I'll try to find the time to provide my list --- at this moment, it
   will be more like a reading list than a textbook TOC. In the future,
   it will be integrated into the E-book I'm working on
   (http://nars.wang.googlepages.com/gti-summary).
  
   Compared to yours, mine will contain less math and algorithms, but
   more psychology and philosophy.
  
   I'd like to see what Richard and others want to propose. We shouldn't
   try to merge them into one wiki page, but several.
  
   Pei
  
  
  
  
  
   On Tue, Mar 25, 2008 at 7:46 PM, Ben Goertzel [EMAIL PROTECTED]
 wrote:
Hi all,
   
 A lot of students email me asking me what to read to get up to
 speed on
  AGI.
   
 So I started a wiki page called Instead of an AGI Textbook,
   
   
 
 http://www.agiri.org/wiki/Instead_of_an_AGI_Textbook#Computational_Linguistics
   
 Unfortunately I did not yet find time to do much but outline a
 table
 of contents there.
   
 So I'm hoping some of you can chip in and fill in some relevant
 hyperlinks on the pages
 I've created ;-)
   
 For those of you too lazy to click the above link, here is the
 introductory note I put on the wiki page:
   
   
 
   
 I've often lamented the fact that there is no advanced undergrad
 level
 textbook for AGI, analogous to what Russell and Norvig is for
 Narrow
 AI.
   
 Unfortunately, I don't have time to write such a textbook, and no
 one
 else with the requisite knowledge and ability seems to have the
 time
 and inclination either.
   
 So, instead of a textbook, I thought it would make sense to outline
 here what the table of contents of such a textbook might look like,
 and to fill in each section within each chapter in this TOC with a
 few
 links to available online resources dealing with the topic of the
 section.
   
 However, all I found time to do today (March 25, 2008) is make the
 TOC. Maybe later I will fill in the links on each section's page,
 or
 maybe by the time I get around it some other folks will have done
 it.
   
 While nowhere near as good as a textbook, I do think this can be a
 valuable resource for those wanting to get up to speed on AGI
 concepts
 and not knowing where to turn to get started. There are some
 available
 AGI bibliographies, but a structured bibliography like this can
 probably be more useful than an unstructured and heterogeneous one.
   
 Naturally my initial TOC represents some of my own biases, but I
 trust
 that by having others help edit it, these biases will ultimately
 come
 out in the wash.
   
 Just to be clear: the idea here is not to present solely AGI
 material.
 Rather the idea is to present material that I think students would
 do
 well to know, if they want to work on AGI. This includes some AGI,
 some narrow AI, some psychology, some neuroscience, some
 mathematics,
 etc.
   
 ***
   
   
 -- Ben
   
   
 --
 Ben Goertzel, PhD
 CEO, Novamente LLC and Biomind LLC
 Director of Research, SIAI
 [EMAIL PROTECTED]
   
 If men cease to believe that they will one day become gods then
 they
 will surely become worms.
 -- Henry Miller
   
 ---
 agi
 Archives: 

Re: [agi] Instead of an AGI Textbook

2008-03-26 Thread Ben Goertzel
Is there some kind of online software that lets a group of people
update a Mind Map
diagram collaboratively, in the manner of a Wiki page?

This would seem critical if a Mind Map is to really be useful for the purpose
you suggest...

-- Ben

On Wed, Mar 26, 2008 at 8:32 AM, Aki Iskandar [EMAIL PROTECTED] wrote:
 Well ... I can take a shot at putting a diagram together.  Making Mind Maps
 is one way I learn any kind of material I want.

 If the topics in the list(s) are descriptive enough, I can take a shot at
 putting such a diagram together.
  It'd be less work to correct it than to make one, right?

 Hey - whatever helps.  For me, it's a win-win.  It would help me, and it
 would help accomplish what you guys are trying to do.

 Let me know,
  ~Aki



 On Tue, Mar 25, 2008 at 10:40 PM, Ben Goertzel [EMAIL PROTECTED] wrote:
  This kind of diagram would certainly be meaningful, but, it would be a
  lot of work to put together, even more so than a traditional TOC ...
 
 
 
 
  On Tue, Mar 25, 2008 at 11:02 PM, Aki Iskandar [EMAIL PROTECTED] wrote:
   Hi Pei -
  
   What about having a tree like diagram that branches out into either:
   - the different paths / approaches to AGI (for instance: NARS,
 Novamente,
   and Richard's, etc.), with suggested readings at those leaves
- area of study, with suggested readings at those leaves
  
   Or possibly, a Mind Map diagram that shows AGI in the middle, with the
   approaches stemming from it, and then either sub fields, or a reading
 list
   and / or collection of links (though the links may become outdated,
 dead).
  
   Point is, would a diagram help map the field - which caters to the
   differing approaches, and which helps those wanting to chart a course to
   their own learning/study ?
  
   Thanks,
   ~Aki
  
  
  
  
On Tue, Mar 25, 2008 at 9:22 PM, Pei Wang [EMAIL PROTECTED]
 wrote:
Ben,
   
It is a good start!
   
Of course everyone else will disagree --- like what Richard did and
I'm going to do. ;-)
   
I'll try to find the time to provide my list --- at this moment, it
will be more like a reading list than a textbook TOC. In the future,
it will be integrated into the E-book I'm working on
(http://nars.wang.googlepages.com/gti-summary).
   
Compared to yours, mine will contain less math and algorithms, but
more psychology and philosophy.
   
I'd like to see what Richard and others want to propose. We shouldn't
try to merge them into one wiki page, but several.
   
Pei
   
   
   
   
   
On Tue, Mar 25, 2008 at 7:46 PM, Ben Goertzel [EMAIL PROTECTED]
 wrote:
 Hi all,

  A lot of students email me asking me what to read to get up to
 speed on
   AGI.

  So I started a wiki page called Instead of an AGI Textbook,


  
 http://www.agiri.org/wiki/Instead_of_an_AGI_Textbook#Computational_Linguistics

  Unfortunately I did not yet find time to do much but outline a
 table
  of contents there.

  So I'm hoping some of you can chip in and fill in some relevant
  hyperlinks on the pages
  I've created ;-)

  For those of you too lazy to click the above link, here is the
  introductory note I put on the wiki page:


  

  I've often lamented the fact that there is no advanced undergrad
 level
  textbook for AGI, analogous to what Russell and Norvig is for
 Narrow
  AI.

  Unfortunately, I don't have time to write such a textbook, and no
 one
  else with the requisite knowledge and ability seems to have the
 time
  and inclination either.

  So, instead of a textbook, I thought it would make sense to outline
  here what the table of contents of such a textbook might look like,
  and to fill in each section within each chapter in this TOC with a
 few
  links to available online resources dealing with the topic of the
  section.

  However, all I found time to do today (March 25, 2008) is make the
  TOC. Maybe later I will fill in the links on each section's page,
 or
  maybe by the time I get around it some other folks will have done
 it.

  While nowhere near as good as a textbook, I do think this can be a
  valuable resource for those wanting to get up to speed on AGI
 concepts
  and not knowing where to turn to get started. There are some
 available
  AGI bibliographies, but a structured bibliography like this can
  probably be more useful than an unstructured and heterogeneous one.

  Naturally my initial TOC represents some of my own biases, but I
 trust
  that by having others help edit it, these biases will ultimately
 come
  out in the wash.

  Just to be clear: the idea here is not to present solely AGI
 material.
  Rather the idea is to present material that I think students would
 do
  well to know, if they want to work on AGI. This includes some AGI,
  some narrow AI, some 

Re: [agi] The Effect of Application of an Idea

2008-03-26 Thread Jim Bromer
On Tue, Mar 25, 2008 at 7:19 PM, Vladimir Nesov [EMAIL PROTECTED] wrote:


 Certainly ambiguity (=applicability to multiple contexts in different
 ways) and presence of rich structure in presumably simple 'ideas', as
 you call it, is a known issue. Even interaction between concept clouds
 evoked by a pair of words is a nontrivial process (triangular
 lightbulb). In a way, whole operation can be modeled by such
 interactions, where sensory input/recall is taken to present a stream
 of triggers that evoke concept cloud after cloud, with associations
 and compound concepts forming at the overlaps. But of course it's too
 hand-wavy without a more restricted model of what's going on.
 Communicating something that exists solely on high level is very
 inefficient, plus most of such content can turn out to be wrong. Back
 to prototyping...

 --
 Vladimir Nesov
 [EMAIL PROTECTED]



I agreed with you up until your conclusion.  While the problems that I
talked about may be known issues, they are discussed almost exclusively
using intuitive models, like we used, or by referring to ineffective models,
like network theories that do not explicitly show how its associative
interrelations would effectively deal with the intricate conceptual details
that would be required to address these issues and would be produced by an
effective solution. I have never seen any theory that was designed to
specifically address the range of situations that I am thinking of
although most earlier AI models were intended to deal similar issues and I
have seen some exemplary models that did use controlled models which showed
how some of these interrelations might be modeled.  These intuitive
discussions and the exaggerated effectiveness of inadequate programs creates
a concept cloud itself, and the problem is that the knowledgeable listener
has a feeling that he understands the problem even without having made any
kind of commitment to the exploration of an effective solution.



Although I have not detailed how the effects of the application of ideas
might be modeled in an actual AI program (or in an extremely simple model
that I would use to start studying the modeling) my whole point is that if
you are interested in advancing AI programming, then the issue that my
theory addresses is a problem that can not be dismissed with a wave of the
hand.  The next step for me is to find a model that would be strong enough
to hold up to genuine extensible learning.



If you are making a decision on how much time you should spend thinking
about this based only on whether or not you have thought about similar
problems I believe that you have already considered some sampling of the
kind of problems that my theory is meant to address.


Jim Bromer

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


Re: [agi] Instead of an AGI Textbook

2008-03-26 Thread Mark Waser

Hi Ben,

   I have a publisher who would love to publish the result of the wiki as a 
textbook if you are willing.


   Mark

- Original Message - 
From: Ben Goertzel [EMAIL PROTECTED]

To: agi@v2.listbox.com
Sent: Tuesday, March 25, 2008 7:46 PM
Subject: [agi] Instead of an AGI Textbook



Hi all,

A lot of students email me asking me what to read to get up to speed on 
AGI.


So I started a wiki page called Instead of an AGI Textbook,

http://www.agiri.org/wiki/Instead_of_an_AGI_Textbook#Computational_Linguistics

Unfortunately I did not yet find time to do much but outline a table
of contents there.

So I'm hoping some of you can chip in and fill in some relevant
hyperlinks on the pages
I've created ;-)

For those of you too lazy to click the above link, here is the
introductory note I put on the wiki page:




I've often lamented the fact that there is no advanced undergrad level
textbook for AGI, analogous to what Russell and Norvig is for Narrow
AI.

Unfortunately, I don't have time to write such a textbook, and no one
else with the requisite knowledge and ability seems to have the time
and inclination either.

So, instead of a textbook, I thought it would make sense to outline
here what the table of contents of such a textbook might look like,
and to fill in each section within each chapter in this TOC with a few
links to available online resources dealing with the topic of the
section.

However, all I found time to do today (March 25, 2008) is make the
TOC. Maybe later I will fill in the links on each section's page, or
maybe by the time I get around it some other folks will have done it.

While nowhere near as good as a textbook, I do think this can be a
valuable resource for those wanting to get up to speed on AGI concepts
and not knowing where to turn to get started. There are some available
AGI bibliographies, but a structured bibliography like this can
probably be more useful than an unstructured and heterogeneous one.

Naturally my initial TOC represents some of my own biases, but I trust
that by having others help edit it, these biases will ultimately come
out in the wash.

Just to be clear: the idea here is not to present solely AGI material.
Rather the idea is to present material that I think students would do
well to know, if they want to work on AGI. This includes some AGI,
some narrow AI, some psychology, some neuroscience, some mathematics,
etc.

***


-- Ben


--
Ben Goertzel, PhD
CEO, Novamente LLC and Biomind LLC
Director of Research, SIAI
[EMAIL PROTECTED]

If men cease to believe that they will one day become gods then they
will surely become worms.
-- Henry Miller

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?;

Powered by Listbox: http://www.listbox.com




---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


Re: [agi] Instead of an AGI Textbook

2008-03-26 Thread Stephen Reed
Ben,
Wikipedia has significant overlap with the topic list on the AGIRI Wiki.  I 
propose for discussion the notion that the AGIRI Wiki be content-compatible 
with Wikipedia along two dimensions:
license - authors agree to the GNU Free Documentation Licenseeditorial 
standards - Wikipedia says that content should be sourced from one or more 
research papers or textbooks, not just from the personal knowledge of the 
author, or from some web page.  I think that this principal should be followed 
to the degree possible.  We could, for example, quarantine non-sourced content 
into clearly marked sections of the containing article.  Typical non-sourced 
content would be some OpenCog API that is not yet published in a journal 
article, technical report, or text book.I concede in advance that most AGIRI 
Wiki authors will find Wikipedia editorial standards burdensome, but the 
benefit would be athat content from the AGIRI Wiki can be used to create new, 
or improve existing Wikipedia articles.  And if we can agree that on the  
easy-to-achieve license, content from Wikipedia, e.g. my article on 
Hierarchical control systems can easily be imported into the AGIRI
 Wiki.

Wikipedia is important to AGI, not only as an online encyclopedia that 
facilitates almost universal access to AGI related topics, but as a target for 
AI researchers that want to structure the text into a vast knowledge base.  
Somewhere down the road to self-improvement, an AGI will be reading Wikipedia.
 
Stephen L. Reed
-Steve

Artificial Intelligence Researcher
http://texai.org/blog
http://texai.org
3008 Oak Crest Ave.
Austin, Texas, USA 78704
512.791.7860

- Original Message 
From: Ben Goertzel [EMAIL PROTECTED]
To: agi@v2.listbox.com
Sent: Tuesday, March 25, 2008 6:46:24 PM
Subject: [agi] Instead of an AGI Textbook

 Hi all,

A lot of students email me asking me what to read to get up to speed on AGI.

So I started a wiki page called Instead of an AGI Textbook,

http://www.agiri.org/wiki/Instead_of_an_AGI_Textbook#Computational_Linguistics

Unfortunately I did not yet find time to do much but outline a table
of contents there.

So I'm hoping some of you can chip in and fill in some relevant
hyperlinks on the pages
I've created ;-)

For those of you too lazy to click the above link, here is the
introductory note I put on the wiki page:




I've often lamented the fact that there is no advanced undergrad level
textbook for AGI, analogous to what Russell and Norvig is for Narrow
AI.

Unfortunately, I don't have time to write such a textbook, and no one
else with the requisite knowledge and ability seems to have the time
and inclination either.

So, instead of a textbook, I thought it would make sense to outline
here what the table of contents of such a textbook might look like,
and to fill in each section within each chapter in this TOC with a few
links to available online resources dealing with the topic of the
section.

However, all I found time to do today (March 25, 2008) is make the
TOC. Maybe later I will fill in the links on each section's page, or
maybe by the time I get around it some other folks will have done it.

While nowhere near as good as a textbook, I do think this can be a
valuable resource for those wanting to get up to speed on AGI concepts
and not knowing where to turn to get started. There are some available
AGI bibliographies, but a structured bibliography like this can
probably be more useful than an unstructured and heterogeneous one.

Naturally my initial TOC represents some of my own biases, but I trust
that by having others help edit it, these biases will ultimately come
out in the wash.

Just to be clear: the idea here is not to present solely AGI material.
Rather the idea is to present material that I think students would do
well to know, if they want to work on AGI. This includes some AGI,
some narrow AI, some psychology, some neuroscience, some mathematics,
etc.

***


-- Ben


-- 
Ben Goertzel, PhD
CEO, Novamente LLC and Biomind LLC
Director of Research, SIAI
[EMAIL PROTECTED]

If men cease to believe that they will one day become gods then they
will surely become worms.
-- Henry Miller

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: http://www.listbox.com/member/?;
Powered by Listbox: http://www.listbox.com







  

Looking for last minute shopping deals?  
Find them fast with Yahoo! Search.  
http://tools.search.yahoo.com/newsearch/category.php?category=shopping

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com

Re: [agi] Instead of an AGI Textbook

2008-03-26 Thread Ben Goertzel
Thanks Mark ... let's see how it evolves...

I think the problem is not finding a publisher, but rather, finding
the time to contribute and refine the content

Maybe in a year or two there will be enough good content there that
someone with appropriate time and inclination and skill can shape it
into a textbook

-- Ben

On Wed, Mar 26, 2008 at 9:49 AM, Mark Waser [EMAIL PROTECTED] wrote:
 Hi Ben,

 I have a publisher who would love to publish the result of the wiki as a
  textbook if you are willing.

 Mark



  - Original Message -
  From: Ben Goertzel [EMAIL PROTECTED]
  To: agi@v2.listbox.com
  Sent: Tuesday, March 25, 2008 7:46 PM
  Subject: [agi] Instead of an AGI Textbook


   Hi all,
  
   A lot of students email me asking me what to read to get up to speed on
   AGI.
  
   So I started a wiki page called Instead of an AGI Textbook,
  
   
 http://www.agiri.org/wiki/Instead_of_an_AGI_Textbook#Computational_Linguistics
  
   Unfortunately I did not yet find time to do much but outline a table
   of contents there.
  
   So I'm hoping some of you can chip in and fill in some relevant
   hyperlinks on the pages
   I've created ;-)
  
   For those of you too lazy to click the above link, here is the
   introductory note I put on the wiki page:
  
  
   
  
   I've often lamented the fact that there is no advanced undergrad level
   textbook for AGI, analogous to what Russell and Norvig is for Narrow
   AI.
  
   Unfortunately, I don't have time to write such a textbook, and no one
   else with the requisite knowledge and ability seems to have the time
   and inclination either.
  
   So, instead of a textbook, I thought it would make sense to outline
   here what the table of contents of such a textbook might look like,
   and to fill in each section within each chapter in this TOC with a few
   links to available online resources dealing with the topic of the
   section.
  
   However, all I found time to do today (March 25, 2008) is make the
   TOC. Maybe later I will fill in the links on each section's page, or
   maybe by the time I get around it some other folks will have done it.
  
   While nowhere near as good as a textbook, I do think this can be a
   valuable resource for those wanting to get up to speed on AGI concepts
   and not knowing where to turn to get started. There are some available
   AGI bibliographies, but a structured bibliography like this can
   probably be more useful than an unstructured and heterogeneous one.
  
   Naturally my initial TOC represents some of my own biases, but I trust
   that by having others help edit it, these biases will ultimately come
   out in the wash.
  
   Just to be clear: the idea here is not to present solely AGI material.
   Rather the idea is to present material that I think students would do
   well to know, if they want to work on AGI. This includes some AGI,
   some narrow AI, some psychology, some neuroscience, some mathematics,
   etc.
  
   ***
  
  
   -- Ben
  
  
   --
   Ben Goertzel, PhD
   CEO, Novamente LLC and Biomind LLC
   Director of Research, SIAI
   [EMAIL PROTECTED]
  
   If men cease to believe that they will one day become gods then they
   will surely become worms.
   -- Henry Miller
  

  ---
   agi
   Archives: http://www.listbox.com/member/archive/303/=now
   RSS Feed: http://www.listbox.com/member/archive/rss/303/
   Modify Your Subscription:
   http://www.listbox.com/member/?;
   Powered by Listbox: http://www.listbox.com
  


  ---
  agi
  Archives: http://www.listbox.com/member/archive/303/=now
  RSS Feed: http://www.listbox.com/member/archive/rss/303/

 Modify Your Subscription: http://www.listbox.com/member/?;


 Powered by Listbox: http://www.listbox.com




-- 
Ben Goertzel, PhD
CEO, Novamente LLC and Biomind LLC
Director of Research, SIAI
[EMAIL PROTECTED]

If men cease to believe that they will one day become gods then they
will surely become worms.
-- Henry Miller

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


Re: [agi] Instead of an AGI Textbook

2008-03-26 Thread Ben Goertzel
Hi Stephen,

 Ben,
 Wikipedia has significant overlap with the topic list on the AGIRI Wiki.  I
 propose for discussion the notion that the AGIRI Wiki be content-compatible
 with Wikipedia along two dimensions:

 license - authors agree to the GNU Free Documentation License

I have no problem with that

 editorial standards - Wikipedia says that content should be sourced from one
 or more research papers or textbooks, not just from the personal knowledge
 of the author, or from some web page.

Well, I think it is appropriate that a wiki covering an in-development research
area should contain a mix of sourced and non-sourced contents, actually.

In many cases it's the non-sourced content that will be the most
valuable, because
it represents practical knowledge and experience of AGI researchers and
developers, which is too new or raw to have been put into the formal literature
yet.

I concede in
 advance that most AGIRI Wiki authors will find Wikipedia editorial standards
 burdensome,

To me this is a pretty major point.

The challenge with an AGI wiki right now is to get people to contribute quality
content at all ... so I'm not psyched about, right now at the starting
stage, making
them jump through hoops in order to do so.

but the benefit would be athat content from the AGIRI Wiki can
 be used to create new, or improve existing Wikipedia articles.

That would be the case so long as the license is in place, it doesn't require
everything to be sourced -- appropriate sourcing could always be
introduced at the time
of porting to Wikipedia.

As the author of a load of academic papers, I'm well aware of how
irritating and
time-consuming it is to properly reference sources.  If I have to do
that for text I place on
the AGIRI wiki, I'm not likely to contribute much to it, just like I
don't currently contribute
much to Wikipedia.  I just don't have the time

And if we
 can agree that on the  easy-to-achieve license, content from Wikipedia, e.g.
 my article on Hierarchical control systems can easily be imported into the
 AGIRI Wiki.

I don't see a problem with the license.

 Wikipedia is important to AGI, not only as an online encyclopedia that
 facilitates almost universal access to AGI related topics, but as a target
 for AI researchers that want to structure the text into a vast knowledge
 base.  Somewhere down the road to self-improvement, an AGI will be reading
 Wikipedia.

Along with the rest of the Web ...  for sure ;-)

-- Ben

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


Re: [agi] Instead of an AGI Textbook

2008-03-26 Thread Stephen Reed
Ben,
I just created an account on the wiki and created my user page derived from my 
Wikipedia user page.  Image uploads on the wiki work the same way as on 
Wikipedia - Yay.
-Steve
 
Stephen L. Reed

Artificial Intelligence Researcher
http://texai.org/blog
http://texai.org
3008 Oak Crest Ave.
Austin, Texas, USA 78704
512.791.7860

- Original Message 
From: Ben Goertzel [EMAIL PROTECTED]
To: agi@v2.listbox.com
Sent: Tuesday, March 25, 2008 6:46:24 PM
Subject: [agi] Instead of an AGI Textbook

 Hi all,

A lot of students email me asking me what to read to get up to speed on AGI.

So I started a wiki page called Instead of an AGI Textbook,

http://www.agiri.org/wiki/Instead_of_an_AGI_Textbook#Computational_Linguistics

Unfortunately I did not yet find time to do much but outline a table
of contents there.

So I'm hoping some of you can chip in and fill in some relevant
hyperlinks on the pages
I've created ;-)

For those of you too lazy to click the above link, here is the
introductory note I put on the wiki page:




I've often lamented the fact that there is no advanced undergrad level
textbook for AGI, analogous to what Russell and Norvig is for Narrow
AI.

Unfortunately, I don't have time to write such a textbook, and no one
else with the requisite knowledge and ability seems to have the time
and inclination either.

So, instead of a textbook, I thought it would make sense to outline
here what the table of contents of such a textbook might look like,
and to fill in each section within each chapter in this TOC with a few
links to available online resources dealing with the topic of the
section.

However, all I found time to do today (March 25, 2008) is make the
TOC. Maybe later I will fill in the links on each section's page, or
maybe by the time I get around it some other folks will have done it.

While nowhere near as good as a textbook, I do think this can be a
valuable resource for those wanting to get up to speed on AGI concepts
and not knowing where to turn to get started. There are some available
AGI bibliographies, but a structured bibliography like this can
probably be more useful than an unstructured and heterogeneous one.

Naturally my initial TOC represents some of my own biases, but I trust
that by having others help edit it, these biases will ultimately come
out in the wash.

Just to be clear: the idea here is not to present solely AGI material.
Rather the idea is to present material that I think students would do
well to know, if they want to work on AGI. This includes some AGI,
some narrow AI, some psychology, some neuroscience, some mathematics,
etc.

***


-- Ben


-- 
Ben Goertzel, PhD
CEO, Novamente LLC and Biomind LLC
Director of Research, SIAI
[EMAIL PROTECTED]

If men cease to believe that they will one day become gods then they
will surely become worms.
-- Henry Miller

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: http://www.listbox.com/member/?;
Powered by Listbox: http://www.listbox.com







  

Never miss a thing.  Make Yahoo your home page. 
http://www.yahoo.com/r/hs

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


Re: [agi] The Effect of Application of an Idea

2008-03-26 Thread Vladimir Nesov
On Wed, Mar 26, 2008 at 4:27 PM, Jim Bromer [EMAIL PROTECTED] wrote:

 I agreed with you up until your conclusion.  While the problems that I
 talked about may be known issues, they are discussed almost exclusively
 using intuitive models, like we used, or by referring to ineffective models,
 like network theories that do not explicitly show how its associative
 interrelations would effectively deal with the intricate conceptual details
 that would be required to address these issues and would be produced by an
 effective solution. I have never seen any theory that was designed to
 specifically address the range of situations that I am thinking of although
 most earlier AI models were intended to deal similar issues and I have seen
 some exemplary models that did use controlled models which showed how some
 of these interrelations might be modeled.  These intuitive discussions and
 the exaggerated effectiveness of inadequate programs creates a concept cloud
 itself, and the problem is that the knowledgeable listener has a feeling
 that he understands the problem even without having made any kind of
 commitment to the exploration of an effective solution.

 Although I have not detailed how the effects of the application of ideas
 might be modeled in an actual AI program (or in an extremely simple model
 that I would use to start studying the modeling) my whole point is that if
 you are interested in advancing AI programming, then the issue that my
 theory addresses is a problem that can not be dismissed with a wave of the
 hand.  The next step for me is to find a model that would be strong enough
 to hold up to genuine extensible learning.

 If you are making a decision on how much time you should spend thinking
 about this based only on whether or not you have thought about similar
 problems I believe that you have already considered some sampling of the
 kind of problems that my theory is meant to address.


What you describe is essentially my own path up to this point: I
started with considering high-level capabilities and gradually worked
towards an implementation that seems to be able to exhibit these
high-level capabilities. At the end of my last message I referred to a
pragmatic problem. Substrate with which I now experiment is
essentially a very simple recurrent network with seemingly
insignificant tweaks. Without high-level view of how to make it
exhibit high-level capabilities I'd never look at it twice. Convincing
someone else that it is that capable will take a rather long
description, and I can well turn out to be wrong (so people have a
perfectly good reason not to listen). It seems more sensible to stick
to prototyping and wait for more solid results, either changing the
theory, or demonstrating its potential.

-- 
Vladimir Nesov
[EMAIL PROTECTED]

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


Re: [agi] Instead of an AGI Textbook

2008-03-26 Thread BillK
On Wed, Mar 26, 2008 at 12:47 PM, Ben Goertzel wrote:
 Is there some kind of online software that lets a group of people
  update a Mind Map diagram collaboratively, in the manner of a Wiki page?

  This would seem critical if a Mind Map is to really be useful for the purpose
  you suggest...



Here is a recent review of online mind mapping software:
http://usableworld.terapad.com/index.cfm?fa=contentNews.newsDetailsnewsID=41870from=listdirectoryId=14375

Online mindmap tools - Updated!
By James Breeze in Mind Maps
Published: Saturday, 08 March 08


BillK

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


Re: [agi] Instead of an AGI Textbook

2008-03-26 Thread Kaj Sotala
On 3/26/08, Ben Goertzel [EMAIL PROTECTED] wrote:
  A lot of students email me asking me what to read to get up to speed on AGI.

Ben,

while we're on the topic, could you elaborate a bit on what kind of
prerequisite knowledge the books you've written/edited require? For
instance, I've been putting off reading Artificial General
Intelligence on the assumption that for the full benefit, it requires
a good understanding of narrow-AI/basic compsci concepts that I
haven't necessarily yet acquired (currently working my way through
Russel  Norvig in order to fix that). The Hidden Pattern sounds like
it would be heavier on the general cogsci/philosophy of mind
requirements, and the Probabilistic Logic Networks one probably needs
a heavy dose of maths (what kind of maths)? What about the
OpenCog/Novamente documentation you've mentioned maybe releasing this
year?

(Agiri.org seems to be down, by the way, so I can't access the textbook page.)


-- 
http://www.saunalahti.fi/~tspro1/ | http://xuenay.livejournal.com/

Organizations worth your time:
http://www.singinst.org/ | http://www.crnano.org/ | http://lifeboat.com/

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


Re: [agi] Instead of an AGI Textbook

2008-03-26 Thread Richard Loosemore



A propos of the several branches of discussion about AGI textbooks on 
this thread...


Knowing what I do about the structure and content of the book I am 
writing, I cannot imagine it being merged as just a set of branch points 
from other works, like the one growing from Ben's TOC.


What I am doing is a coherent body of thought in its own right, with a 
radically different underlying philosophy, so it really needs to be a 
standalone project.




Richard Loosemore

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


Re: [agi] The Effect of Application of an Idea

2008-03-26 Thread William Pearson
On 25/03/2008, Vladimir Nesov [EMAIL PROTECTED] wrote

 Simple systems can be computationally universal, so it's not an issue
  in itself. On the other hand, no learning algorithm is universal,
  there are always distributions that given algorithms will learn
  miserably. The problem is to find a learning algorithm/representation
  that has the right kind of bias to implement human-like performance.

First a riddle: What can be all learning algorithms, but is none?

I'd disagree. Okay simple systems can be computationally universal,
but what does that really mean.

Computational universality means to be able to represent any
computable function, the range and domain of this function are assumed
to be from the natural numbers to itself.

Most AI formulations when they say that are computationally universal
are only talking about function of F: I → O where I is the input and O
is the output. These include the formulations of neural networks/GA
etc that I have seen. However there are lots of interesting programs
in computers that do not map the input to the output. Humans also do
not just map the input to the output, we also think, ruminate, model
and remember. This does not affect the range of functions from the
input to the output, but it does change how quickly they can be moved
between. What I am interested in is in systems where the ranges and
domains of the functions are entities inside the system.

That is the F: I → S, F: S → O, and F: S→ S are important and should
be potentially computationally universal. Where S is the internal
memory of the system. This allows the system to be all possible
learning algorithms (although only one at any time), but also it is no
algorithm (else F: I x S → S, would be fixed).

General purpose desktop computers are these kinds of systems. If they
weren't how else could we implement any type of learning system on
them? Thus the answer to my riddle.

The question I have been trying to answer precisely is how to govern
these sorts of systems so they roughly do what you want, without you
having to give precise instructions.

  Will Pearson

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


Re: [agi] The Effect of Application of an Idea

2008-03-26 Thread Jim Bromer
On Wed, Mar 26, 2008 at 10:17 AM, Vladimir Nesov [EMAIL PROTECTED] wrote:


 What you describe is essentially my own path up to this point: I
 started with considering high-level capabilities and gradually worked
 towards an implementation that seems to be able to exhibit these
 high-level capabilities. At the end of my last message I referred to a
 pragmatic problem. Substrate with which I now experiment is
 essentially a very simple recurrent network with seemingly
 insignificant tweaks. Without high-level view of how to make it
 exhibit high-level capabilities I'd never look at it twice. Convincing
 someone else that it is that capable will take a rather long
 description, and I can well turn out to be wrong (so people have a
 perfectly good reason not to listen). It seems more sensible to stick
 to prototyping and wait for more solid results, either changing the
 theory, or demonstrating its potential.

 --
 Vladimir Nesov

I do not know much about neural networks, but from  what I read, I always
felt that a recurrent network would be the only way you could feasibly get
an ANN to represent (excuse my french) distinct items without absurdly huge
and noisy expansions. So I am curious about what you are talking about.
When you mention prototyping, you are talking about prototyping the neural
network with high level concepts for easier demonstrations or something like
that.  I think there was some discussion about using 'labels' in neural
networks on one of those links to an online video that were recently
posted.  Is this similar to what you mean by prototyping?
Jim Bromer

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


Re: [agi] The Effect of Application of an Idea

2008-03-26 Thread Jim Bromer
2008/3/26 William Pearson [EMAIL PROTECTED]:

 On 25/03/2008, Vladimir Nesov [EMAIL PROTECTED] wrote
 
  Simple systems can be computationally universal, so it's not an issue
   in itself. On the other hand, no learning algorithm is universal,
   there are always distributions that given algorithms will learn
   miserably. The problem is to find a learning algorithm/representation
   that has the right kind of bias to implement human-like performance.

 First a riddle: What can be all learning algorithms, but is none?


Excellent philosophical point!!


 Okay simple systems can be computationally universal,
 but what does that really mean.

 Computational universality means to be able to represent any
 computable function, the range and domain of this function are assumed
 to be from the natural numbers to itself.


---I think Godel would disagree.


 Most AI formulations when they say that are computationally universal
 are only talking about function of F: I → O where I is the input and O
 is the output. These include the formulations of neural networks/GA
 etc that I have seen. However there are lots of interesting programs
 in computers that do not map the input to the output. Humans also do
 not just map the input to the output, we also think, ruminate, model
 and remember. This does not affect the range of functions from the
 input to the output, but it does change how quickly they can be moved
 between. What I am interested in is in systems where the ranges and
 domains of the functions are entities inside the system.

 That is the F: I → S, F: S → O, and F: S→ S are important and should
 be potentially computationally universal. Where S is the internal
 memory of the system. This allows the system to be all possible
 learning algorithms (although only one at any time), but also it is no
 algorithm (else F: I x S → S, would be fixed).

 General purpose desktop computers are these kinds of systems. If they
 weren't how else could we implement any type of learning system on
 them? Thus the answer to my riddle.

 The question I have been trying to answer precisely is how to govern
 these sorts of systems so they roughly do what you want, without you
 having to give precise instructions.

  Will Pearson


-I am going to read this more carefully later.  However, the first part
of the answer to your last question is that the governance of these kinds of
systems will be based on general rules (or methods of generality) so you do
not need to define all the precise instructions that would be needed.  But,
there is not one level of universality, there are potentially infinite
levels of generalization, and they do not all mesh together perfectly.
Although this kind of talk may not solve the problem, I believe that this is
where we are going to end up working if we continue to work on the problem
Jim Bromer

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


Re: [agi] Instead of an AGI Textbook

2008-03-26 Thread Ben Goertzel
Fair enough, Richard...

Again I'll emphasize that the idea of the Instead of an AGI Textbook
is not to teach any particular theory or design for AGI, but rather to convey
background knowledge that is useful for folks who wish to come to grips
with contemporary AGI theories and designs

I have articulated my own coherent body of thought regarding AGI as well,
but I consider it to best be presented at the research treatise or research
paper rather than textbook level...

-- Ben G


On Wed, Mar 26, 2008 at 12:55 PM, Richard Loosemore [EMAIL PROTECTED] wrote:


  A propos of the several branches of discussion about AGI textbooks on
  this thread...

  Knowing what I do about the structure and content of the book I am
  writing, I cannot imagine it being merged as just a set of branch points
  from other works, like the one growing from Ben's TOC.

  What I am doing is a coherent body of thought in its own right, with a
  radically different underlying philosophy, so it really needs to be a
  standalone project.



  Richard Loosemore



  ---
  agi
  Archives: http://www.listbox.com/member/archive/303/=now
  RSS Feed: http://www.listbox.com/member/archive/rss/303/
  Modify Your Subscription: http://www.listbox.com/member/?;
  Powered by Listbox: http://www.listbox.com




-- 
Ben Goertzel, PhD
CEO, Novamente LLC and Biomind LLC
Director of Research, SIAI
[EMAIL PROTECTED]

If men cease to believe that they will one day become gods then they
will surely become worms.
-- Henry Miller

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


Re: [agi] The Effect of Application of an Idea

2008-03-26 Thread Vladimir Nesov
On Wed, Mar 26, 2008 at 8:47 PM, Jim Bromer [EMAIL PROTECTED] wrote:

 I do not know much about neural networks, but from  what I read, I always
 felt that a recurrent network would be the only way you could feasibly get
 an ANN to represent (excuse my french) distinct items without absurdly huge
 and noisy expansions. So I am curious about what you are talking about.
 When you mention prototyping, you are talking about prototyping the neural
 network with high level concepts for easier demonstrations or something like
 that.  I think there was some discussion about using 'labels' in neural
 networks on one of those links to an online video that were recently posted.
 Is this similar to what you mean by prototyping?


For now objective is to try to achieve basic high-level dynamics that
this architecture was designed to implement, and thus to partially
establish consistence of many-faceted high-level design with simple
network implementation. If this stage succeeds, after a bit of
scalability implementation it should be possible to start teaching it
more impressive high-level feats.

-- 
Vladimir Nesov
[EMAIL PROTECTED]

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


Re: [agi] Microsoft Launches Singularity

2008-03-26 Thread Charles D Hixson

John G. Rose wrote:

From: Richard Loosemore [mailto:[EMAIL PROTECTED]
My take on this is completely different.

When I say Narrow AI I am specifically referring to something that is
so limited that it has virtually no chance of becoming a general
intelligence.  There is more to general intelligence than just throwing
a bunch of Narrow AI ideas into a pot and hoping for the best. If it
were, we would have had AGI long before now.



It's an opinion that AGI could not be built out of a conglomeration of
narrow-AI subcomponents. Also there are many things that COULD be built with
narrow-AI that we have not even scratched the surface of due to a number of
different limitations so saying that we would have achieved AGI long ago is
an exaggeration.
  
I don't think a General Intelligence could be built entirely out of 
narrow AI components, but it might well be a relatively trivial add-on.  
Just consider how much of human intelligence is demonstrably narrow AI 
(well, not artificial, but you know what I mean).  Object recognition, 
e.g.  Then start trying to guess how much of the part that we can't 
prove a classification for is likely to be a narrow intelligence 
component.  In my estimation (without factual backing) less than 0.001 
of our intelligence is General Intellignece, possibly much less.
 
  

Consciousness and self-awareness are things that come as part of the AGI
package.  If the system is too simple to have/do these things, it will
not be general enough to equal the human mind.




I feel that general intelligence may not require consciousness and
self-awareness. I am not sure of this and may prove myself wrong. To equal
the human mind you need these things of course and to satisfy the sci-fi
fantasy world's appetite for intelligent computers you would need to
incorporate these as well.

John
  
I'm not sure of the distinction that you are making between 
consciousness and self-awareness, but even most complex narrow-AI 
applications require at least rudimentary self awareness.  In fact, one 
could argue that all object oriented programming with inheritance has 
rudimentary self awareness (called this in many languages, but in 
others called self).  This may be too rudimentary, but it's my feeling 
that it's an actual model(implementation?) of what the concept of self 
has evolved from.


As to an AGI not being conscious I'd need to see a definition of 
your terms, because otherwise I've *got* to presume that we have 
radically different definitions.  To me an AGI would not only need to be 
aware of itself, but also to be aware of aspects of it's environment 
that it could effect changes in,  And of the difference between them, 
though that might well be learned.  (Zen:  Who is the master who makes 
the grass green?, and a few other koans when solved imply that in 
humans the distinction between internal and external is a learned 
response.)  Perhaps the diagnostic characteristic of an AGI is that it 
CAN learn that kind of thing.  Perhaps not, too.  I can imagine a narrow 
AI that was designed to plug into different bodies, and in each case 
learn the distinction between itself and the environment before 
proceeding with its assignment.  I'm not sure it's possible, but I can 
imagine it.


OTOH, if we take my arguments in the preceding paragraph too seriously, 
then medical patients that are locked in would be considered not 
intelligent.  This is clearly incorrect.  Effectively they aren't 
intelligent, but that's because of a mechanical breakdown in the 
sensory/motor area, and that clearly isn't what we mean when we talk 
about intelligence.  But examples of recovered/recovering patients seem 
to imply that they weren't exactly either intelligent or conscious while 
they were locked-in.  (I'm going solely by reports in the popular 
science press...so don't take this too seriously.)  It appears as if 
when external sensations are cut-off, that the mind estivates...at least 
after awhile.  Presumably different patients had different causes, and 
thence at least slightly different effects, but that's my first-cut 
guess at what's happening.  OTOH, the sensory/motor channel doesn't need 
to be particularly well functioning.  Look at Stephan Hawking.


---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


[agi] Why Hugo de Garis is WRONG!

2008-03-26 Thread Mark Waser
Before swatting at one of those pesky flies that come out as the days lengthen 
and the temperature rises, one should probably think twice. A University of 
Missouri researcher has found, through the study of Drosophila (a type of fruit 
fly), that by manipulating levels of certain compounds associated with the 
circuitry of the brain, key genes related to memory can be isolated and 
tested. The results of the study may benefit human patients suffering from 
Parkinson's disease and could eventually lead to discoveries in the treatment 
of depression.

http://www.machineslikeus.com/cms/news/flys-small-brain-may-benefit-humans

Mark

Vision/Slogan -- Friendliness:  The Ice-9 of Ethics and Ultimate in 
Self-Interest

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


Fwd: [agi] Instead of an AGI Textbook

2008-03-26 Thread Ben Goertzel
 BTW I improved the hierarchical organization of the TOC a bit, to
 remove the impression that it's just a random grab-bag of topics...


 http://www.agiri.org/wiki/Instead_of_an_AGI_Textbook

 ben

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


[agi] META: email format (was Why Hugo de Garis is WRONG!)

2008-03-26 Thread J. Andrew Rogers

Hi Mark,

Could you *please* not send HTML email?  Ignoring that it is generally  
considered poor netiquette, and for good reason, it frequently gets  
turned into barely readable hash by even the most modern email clients.


I am using Mail.app 2.0 on OSX 10.5 which handles rendering better  
than most, and most HTML email is *still* generally rendered as far  
uglier and less readable than plaintext email.  Given that HTML email  
does not add anything substantive could we please stick to plaintext  
for the sake of communication?


Thanks,

J. Andrew Rogers


On Mar 26, 2008, at 11:37 AM, Mark Waser wrote:
Before swatting at one of those pesky flies that come out as the  
days lengthen and the temperature rises, one should probably think  
twice. A University of Missouri researcher has found, through the  
study of Drosophila (a type of fruit fly), that by manipulating  
levels of certain compounds associated with the circuitry of the  
brain, key genes related to memory can be isolated and tested. The  
results of the study may benefit human patients suffering from  
Parkinson's disease and could eventually lead to discoveries in the  
treatment of depression.


http://www.machineslikeus.com/cms/news/flys-small-brain-may-benefit-humans

Mark

Vision/Slogan -- Friendliness:  The Ice-9 of Ethics and Ultimate in  
Self-Interest

agi | Archives  | Modify Your Subscription



---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


Re: Mark Launches Singularity :-) WAS Re: [agi] Microsoft Launches Singularity

2008-03-26 Thread Mark Waser
:-)  Now that is *funny* -- and polite -- especially after said ASCII diagrams 
got so badly mangled (and I got flamed for using HTML e-mail -- which is even 
more humorous ;-)
  - Original Message - 
  From: Chris Petersen 
  To: agi@v2.listbox.com 
  Sent: Wednesday, March 26, 2008 3:52 AM
  Subject: Re: Mark Launches Singularity :-) WAS Re: [agi] Microsoft Launches 
Singularity


  Mentifex called; it wants its ASCII diagrams back.

  -Chris



--
agi | Archives  | Modify Your Subscription  

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


Re: [agi] META: email format (was Why Hugo de Garis is WRONG!)

2008-03-26 Thread Charles D Hixson

J. Andrew Rogers wrote:

Hi Mark,

Could you *please* not send HTML email?  Ignoring that it is generally 
considered poor netiquette, and for good reason, it frequently gets 
turned into barely readable hash by even the most modern email clients.


I am using Mail.app 2.0 on OSX 10.5 which handles rendering better 
than most, and most HTML email is *still* generally rendered as far 
uglier and less readable than plaintext email.  Given that HTML email 
does not add anything substantive could we please stick to plaintext 
for the sake of communication?


Thanks,

J. Andrew Rogers


On Mar 26, 2008, at 11:37 AM, Mark Waser wrote:
Before swatting at one of those pesky flies that come out as the days 
lengthen and the temperature rises, one should probably think twice. 
A University of Missouri researcher has found, through the study of 
Drosophila (a type of fruit fly), that by manipulating levels of 
certain compounds associated with the circuitry of the brain, key 
genes related to memory can be isolated and tested. The results of 
the study may benefit human patients suffering from Parkinson's 
disease and could eventually lead to discoveries in the treatment of 
depression.


http://www.machineslikeus.com/cms/news/flys-small-brain-may-benefit-humans 



Mark

Vision/Slogan -- Friendliness:  The Ice-9 of Ethics and Ultimate in 
Self-Interest

agi | Archives  | Modify Your Subscription

I suspect that he's using a webmail client of some sort.  The body of 
his message is, essentially, text.  It's only a couple of buttons(?) in 
the footer that are clearly intentionally html, and he probably doesn't 
add them.  So my guess is that his mail client is wrapping his e-mail in 
an html frame of some sort and sticking, probably, advertisements at the 
bottom.  (I screen remote images out of e-mail, so I don't really know.


This is the part I'm talking about:

table border=3D0 cellspacing=3D0 cellpadding=3D0 width=3D100% styl=
e=3Dbackground-color:#fff bgcolor=3D#ff
 tr
   td padding=3D4px
 font color=3Dblack size=3D1 face=3Dhelvetica, sans-serif;
 strongagi/strong | a style=3Dtext-decoration:none;color:#669933=
;border-bottom: 1px solid #44
href=3Dhttp://www.listbox.com/member/archive/303/=3Dnow; title=3DGo to ar=
chives for agiArchives/a
a border=3D0 style=3Dtext-decoration:none;color:#669933 href=3Dhttp:/=
/www.listbox.com/member/archive/rss/303/ title=3DRSS feed for agiimg b=
order=3D0 src=3Dhttps://www.listbox.com/images/feed-icon-10x10.jpg;/a
| a style=3Dtext-decoration:none;color:#669933;border-bottom: 1px solid =
#44
href=3Dhttp://www.listbox.com/member/?D232072D98557=
868-5cf207 title=3DModify/a
Your Subscriptiontd valign=3Dtop align=3Drighta style=3Dborder-bot=
tom:none; href=3Dhttp://www.listbox.com;

img src=3Dhttps://www.listbox.com/images/listbox-logo-small.jpg;
title=3DPowered by Listbox border=3D0 //a/td

That said, I agree with you about it's inadvisability.  But often the sender 
isn't even aware of what impression he is making.
I've given up on refusing to accept html e-mail.  Too many people don't even 
know what you're talking about.  But I definitely won't accept remote images or 
such.  Such things are dangerous.

 /font
   /td
 /tr
/table


---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


I know, I KNOW :-) WAS Re: [agi] The Effect of Application of an Idea

2008-03-26 Thread Mark Waser

First a riddle: What can be all learning algorithms, but is none?


A human being!


---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


Re: I know, I KNOW :-) WAS Re: [agi] The Effect of Application of an Idea

2008-03-26 Thread William Pearson
On 26/03/2008, Mark Waser [EMAIL PROTECTED] wrote:
  First a riddle: What can be all learning algorithms, but is none?

  A human being!


Well my answer was a common PC, which I hope is more illuminating
because we know it well.

But human being works, as does any future AI design, as far as I am concerned.

  Will Pearson

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


[agi] Instead_of_an_AGI_Textbook Challenge !!

2008-03-26 Thread Ben Goertzel
OK... I just burned an hour inserting more links and content into

http://www.agiri.org/wiki/Instead_of_an_AGI_Textbook

I'm burnt out on it for a while, there's too much other stuff on my plate

However, I have a challenge for y'all

There are something like 400 people subscribed to this list...

If 25 of you spend 30 minutes each, during the next week, adding
relevant content to the non-textbook wiki page ... then at the end
of the week we will have a pretty nice knowledge resource for
newbies to AGI.

And we will probably all learn something from following up each
others' references ...

And then I'll save a lot of time during the next year, because when
someone emails me and asks me what they should read to get
up to speed on the general thinking in the AGI field, I'll just point
them to the non-textbook ;-)

-- Ben




-- 
Ben Goertzel, PhD
CEO, Novamente LLC and Biomind LLC
Director of Research, SIAI
[EMAIL PROTECTED]

If men cease to believe that they will one day become gods then they
will surely become worms.
-- Henry Miller

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


[agi] Re: Instead_of_an_AGI_Textbook Challenge !!

2008-03-26 Thread Ben Goertzel
Ah, one more note...

Due to its location on the AGIRI wiki, the Instead_of_an_AGI_Textbook
automatically links into the Mind Ontology

http://www.agiri.org/wiki/Mind_Ontology

that I created in a fit of mania one weekend a couple years ago.

So, just remember that if you decide to add content to the non-textbook,
rather than just links, you can link it into the Mind Ontology, expand
the Mind Ontology, etc.

The idea of the Mind Ontology was to create a unified common vocabulary
for AGI thinkers/researchers...

It didn't really work because almost no one paid attention, but it was a sort
of fun weekend ;-)

-- Ben


On Wed, Mar 26, 2008 at 10:43 PM, Ben Goertzel [EMAIL PROTECTED] wrote:
 OK... I just burned an hour inserting more links and content into

  http://www.agiri.org/wiki/Instead_of_an_AGI_Textbook

  I'm burnt out on it for a while, there's too much other stuff on my plate

  However, I have a challenge for y'all

  There are something like 400 people subscribed to this list...

  If 25 of you spend 30 minutes each, during the next week, adding
  relevant content to the non-textbook wiki page ... then at the end
  of the week we will have a pretty nice knowledge resource for
  newbies to AGI.

  And we will probably all learn something from following up each
  others' references ...

  And then I'll save a lot of time during the next year, because when
  someone emails me and asks me what they should read to get
  up to speed on the general thinking in the AGI field, I'll just point
  them to the non-textbook ;-)

  -- Ben




  --
  Ben Goertzel, PhD
  CEO, Novamente LLC and Biomind LLC
  Director of Research, SIAI
  [EMAIL PROTECTED]

  If men cease to believe that they will one day become gods then they
  will surely become worms.
  -- Henry Miller




-- 
Ben Goertzel, PhD
CEO, Novamente LLC and Biomind LLC
Director of Research, SIAI
[EMAIL PROTECTED]

If men cease to believe that they will one day become gods then they
will surely become worms.
-- Henry Miller

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


Re: [agi] Re: Instead_of_an_AGI_Textbook Challenge !!

2008-03-26 Thread Matthew Henderson
Hey Ben, I really appreciate this idea.  I recently became very interested
in the Singularity, AI, etc.  I very much enjoy the discussions that go on
in these threads, but rarely know what is going on.  I'm really hoping to
learn everything there is to know from the ground up.

Matt

On Wed, Mar 26, 2008 at 9:46 PM, Ben Goertzel [EMAIL PROTECTED] wrote:

 Ah, one more note...

 Due to its location on the AGIRI wiki, the Instead_of_an_AGI_Textbook
 automatically links into the Mind Ontology

 http://www.agiri.org/wiki/Mind_Ontology

 that I created in a fit of mania one weekend a couple years ago.

 So, just remember that if you decide to add content to the non-textbook,
 rather than just links, you can link it into the Mind Ontology, expand
 the Mind Ontology, etc.

 The idea of the Mind Ontology was to create a unified common vocabulary
 for AGI thinkers/researchers...

 It didn't really work because almost no one paid attention, but it was a
 sort
 of fun weekend ;-)

 -- Ben


 On Wed, Mar 26, 2008 at 10:43 PM, Ben Goertzel [EMAIL PROTECTED] wrote:
  OK... I just burned an hour inserting more links and content into
 
   http://www.agiri.org/wiki/Instead_of_an_AGI_Textbook
 
   I'm burnt out on it for a while, there's too much other stuff on my
 plate
 
   However, I have a challenge for y'all
 
   There are something like 400 people subscribed to this list...
 
   If 25 of you spend 30 minutes each, during the next week, adding
   relevant content to the non-textbook wiki page ... then at the end
   of the week we will have a pretty nice knowledge resource for
   newbies to AGI.
 
   And we will probably all learn something from following up each
   others' references ...
 
   And then I'll save a lot of time during the next year, because when
   someone emails me and asks me what they should read to get
   up to speed on the general thinking in the AGI field, I'll just point
   them to the non-textbook ;-)
 
   -- Ben
 
 
 
 
   --
   Ben Goertzel, PhD
   CEO, Novamente LLC and Biomind LLC
   Director of Research, SIAI
   [EMAIL PROTECTED]
 
   If men cease to believe that they will one day become gods then they
   will surely become worms.
   -- Henry Miller
 



 --
 Ben Goertzel, PhD
 CEO, Novamente LLC and Biomind LLC
 Director of Research, SIAI
 [EMAIL PROTECTED]

 If men cease to believe that they will one day become gods then they
 will surely become worms.
 -- Henry Miller

 ---
 agi
 Archives: http://www.listbox.com/member/archive/303/=now
 RSS Feed: http://www.listbox.com/member/archive/rss/303/
 Modify Your Subscription:
 http://www.listbox.com/member/?;
 Powered by Listbox: http://www.listbox.com


---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


Re: [agi] Instead_of_an_AGI_Textbook Challenge !!

2008-03-26 Thread Stephen Reed
Thanks Ben for leaving a placeholder for Fluid Construction Grammar.  I've 
copied over the Wikipedia article for which I wrote most of the content.

Cheers.
-Steve
 
Stephen L. Reed

Artificial Intelligence Researcher
http://texai.org/blog
http://texai.org
3008 Oak Crest Ave.
Austin, Texas, USA 78704
512.791.7860

- Original Message 
From: Ben Goertzel [EMAIL PROTECTED]
To: agi@v2.listbox.com
Sent: Wednesday, March 26, 2008 9:43:35 PM
Subject: [agi] Instead_of_an_AGI_Textbook Challenge !!

OK... I just burned an hour inserting more links and content into

http://www.agiri.org/wiki/Instead_of_an_AGI_Textbook

I'm burnt out on it for a while, there's too much other stuff on my plate

However, I have a challenge for y'all

There are something like 400 people subscribed to this list...

If 25 of you spend 30 minutes each, during the next week, adding
relevant content to the non-textbook wiki page ... then at the end
of the week we will have a pretty nice knowledge resource for
newbies to AGI.

And we will probably all learn something from following up each
others' references ...

And then I'll save a lot of time during the next year, because when
someone emails me and asks me what they should read to get
up to speed on the general thinking in the AGI field, I'll just point
them to the non-textbook ;-)

-- Ben




-- 
Ben Goertzel, PhD
CEO, Novamente LLC and Biomind LLC
Director of Research, SIAI
[EMAIL PROTECTED]

If men cease to believe that they will one day become gods then they
will surely become worms.
-- Henry Miller

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: http://www.listbox.com/member/?;
Powered by Listbox: http://www.listbox.com







  

Looking for last minute shopping deals?  
Find them fast with Yahoo! Search.  
http://tools.search.yahoo.com/newsearch/category.php?category=shopping

---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com


RE: [agi] Microsoft Launches Singularity

2008-03-26 Thread John G. Rose
 From: Charles D Hixson [mailto:[EMAIL PROTECTED]
 
 I don't think a General Intelligence could be built entirely out of
 narrow AI components, but it might well be a relatively trivial add-on.
 Just consider how much of human intelligence is demonstrably narrow AI
 (well, not artificial, but you know what I mean).  Object recognition,
 e.g.  Then start trying to guess how much of the part that we can't
 prove a classification for is likely to be a narrow intelligence
 component.  In my estimation (without factual backing) less than 0.001
 of our intelligence is General Intellignece, possibly much less.
 

I agree that it may be 1%. Also from what I've read with brain atrophy
cases is that a typical human brain may be able to function relatively
normally with 10% of its mass if atrophy is applied over time.

 I'm not sure of the distinction that you are making between
 consciousness and self-awareness, but even most complex narrow-AI
 applications require at least rudimentary self awareness.  In fact, one
 could argue that all object oriented programming with inheritance has
 rudimentary self awareness (called this in many languages, but in
 others called self).  This may be too rudimentary, but it's my feeling
 that it's an actual model(implementation?) of what the concept of self
 has evolved from.
 

Consciousness and awareness are two functions that I was separating out. The
programming language this and self are particular to class instances
right and can be at the root of the hierarchy tree but there are many, many
this's in a large OO application. A collective group could be considered
some sort of self-awareness this is true and it could be fleshed out and
expanded upon. What I have been exploring though is whether conscious,
awareness, etc. have to be present for a general intelligence. The trend is
to include them.

 As to an AGI not being conscious I'd need to see a definition of
 your terms, because otherwise I've *got* to presume that we have
 radically different definitions.  To me an AGI would not only need to be
 aware of itself, but also to be aware of aspects of it's environment
 that it could effect changes in,  And of the difference between them,
 though that might well be learned.  (Zen:  Who is the master who makes
 the grass green?, and a few other koans when solved imply that in
 humans the distinction between internal and external is a learned
 response.)  Perhaps the diagnostic characteristic of an AGI is that it
 CAN learn that kind of thing.  Perhaps not, too.  I can imagine a narrow
 AI that was designed to plug into different bodies, and in each case
 learn the distinction between itself and the environment before
 proceeding with its assignment.  I'm not sure it's possible, but I can
 imagine it.

AGI per se may be defined as a lifelike intelligent entity requiring brain
related things like consciousness. In my mind, I am thinking of general
intelligence without the difficult task of building consciousness. You could
argue a rock has some sort of consciousness. I'm thinking intelligence is a
sort of self-contained entity that depends upon the state, structure,
complexity and potential of its contained data and representation.
Intelligence would be related to an energy transfer needed to extract a
structured data set from a structured data superset. The structured data set
(a query) would have a morphic chain relationship to the structure of the
stored data and the energy required to get it would be proportional to the
intelligence. Lower energy expenditure across query types implies higher
intelligence related to those queries. The morphic chain relationship
basically is a subset of a morphism mapping graph. Better intelligence means
solving the graph and applying optimizing techniques based on parameters.
Measurement of intelligence (the energy) would basically be counting bit
flips on queries related to query structure and bit count. Knowledge
optimization such as self organizing and optimizing morphism graphs
naturally affect the potential energy and things like having this reorganize
based on query is all part of it. But from what I gather intelligence is
just a bit and time (or state) relationship between sets of bits - that is
for a digital based intelligence. I don't know if an analog based
intelligence would have similar mathematical structure or not...I suppose
that when you boil it down they'll be particle wave duality issues :)

John


---
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244id_secret=98558129-0bdb63
Powered by Listbox: http://www.listbox.com