Derek wrote: >>Building computer systems that generate pictures or videos as their way of >>communicating with us could be a very lucrative addition to computer >>applications that include cognitive models of their users (instead of >>focusing solely on generating natural language), because most of us do >>process visual information so well. ------------------------------------------------------------------------------------------- 1. Derek, A simple movie, like a visual slide show or cartoon, would be a fairly straight forward implementation for certain symbolic/connectionist and other models. The architecture I work on is set to do that as a medium term goal. Any representation system (language) is learned within the contextual patterns its used, so a visual language (shapes), motion language (ultimately gestures), text, sound (un-implemented), etc. of all sensor type can be integrated with the core system as interchangeable patterns.* You can simply read off the state from any region to see what it's thinking within specific and multimodal sensors, as well as what it's attention is on and the focus level of that atttention; e.g. looking at the colorful red and blue car in the parking lot, but narrowly focus on the rust on the door handle or a wide defocus on the entire parking lot and the heat from the pavement - all in the same situation. When this situation is recalled or referenced, the scene is pulled-up and it will generate its version of what it was focused on (rust on the door handle or heat). It can then shift it's focus outward or to other regions in the scene, modify a feature (say change the red colors to blue on the car etc). A movie from its output would look like the jerky camera movements and quick focus change in a show like "NYPD Blue" or the new "Battlestar Galactic" space scenes or History Channel's "Dogfight" sky scenes. *pattens (for the model) have spatial and serial components to varying degrees for each sensor: e.g. bodyspace maps easy to vision, sound to change (motion sequence), force (touch) to vision,...for the sensors that don't map well, just force it and you get your less seemless metaphors : a sweet sound, but left the room with a bitter taste
2. Derek wrote: >>....because most of us do process visual information so well. It makes it easier to program visually to see what the system is doingas a whole or by states --- On Wed, 11/26/08, Derek Zahn <[EMAIL PROTECTED]> wrote: From: Derek Zahn <[EMAIL PROTECTED]> Subject: RE: [agi] The Future of AGI To: agi@v2.listbox.com Date: Wednesday, November 26, 2008, 11:02 AM #yiv2133712726 .hmmessage P { margin:0px;padding:0px;} #yiv2133712726 { font-size:10pt;font-family:Verdana;} Although a lot of AI-type research focuses on natural language interfaces between computer systems and their human users, computers have the ability to create visual images (which people can't do in real-time beyond gestures and facial expressions). Building computer systems that generate pictures or videos as their way of communicating with us could be a very lucrative addition to computer applications that include cognitive models of their users (instead of focusing solely on generating natural language), because most of us do process visual information so well. This is really narrow AI I suppose, though it's kind of on the borderline. It does seem like one of the ways to commercialize incremental progress toward AGI. Derek Zahn supermodelling.net agi | Archives | Modify Your Subscription ------------------------------------------- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=120640061-aded06 Powered by Listbox: http://www.listbox.com