--- On Sun, 9/7/08, John G. Rose <[EMAIL PROTECTED]> wrote: > From: John G. Rose <[EMAIL PROTECTED]> > Subject: RE: Language modeling (was Re: [agi] draft for comment) > To: [email protected] > Date: Sunday, September 7, 2008, 9:15 AM > > From: Matt Mahoney [mailto:[EMAIL PROTECTED] > > > > --- On Sat, 9/6/08, John G. Rose > <[EMAIL PROTECTED]> wrote: > > > > > Compression in itself has the overriding goal of > reducing > > > storage bits. > > > > Not the way I use it. The goal is to predict what the > environment will > > do next. Lossless compression is a way of measuring > how well we are > > doing. > > > > Predicting the environment in order to determine which data > to pack where, > thus achieving higher compression ratio. Or compression as > an integral part > of prediction? Some types of prediction are inherently > compressed I suppose.
Predicting the environment to maximize reward. Hutter proved that universal intelligence is a compression problem. The optimal behavior of an AIXI agent is to guess the shortest program consistent with observation so far. That's algorithmic compression. -- Matt Mahoney, [EMAIL PROTECTED] ------------------------------------------- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=111637683-c8fa51 Powered by Listbox: http://www.listbox.com
