Hey Paige Thank you very much for your prompt reply and clarification.
According to my employer, I have only one more month of Medical Leave remaining so I would be able to participate in Arm 11 of the alternative study which I believe consists of radiation only tx for approximately two months, assuming I meet the necessary criteria, and the randomization places me in that ARM, and the tx begins very soon. Otherwise I am afraid I must, regretfully, decline participation in either study. Peter Sent from my iPad > On May 11, 2015, at 12:48 PM, "Greg Staskowski" <[email protected]> wrote: > > Look, guys and Ben. This is all fun and cute, but I'm wondering if we can't > think even bigger? > > For example, is it possible to construct an irrational lambda calculus that > can be paired with our current lambda calculus and modal logic to create > agents that will evolve more towards what I call "emergent silicon organisms" > or ESO's? > > If I'm all wet, go ahead and say so, I don't bruise easily. > > -GJS > >> On Sun, May 10, 2015 at 5:25 PM, Boris Kazachenko <[email protected]> wrote: >> OK, but in general terms, that's what DL is doing anyway. You want to the >> same thing without its flaws, but how? >> You know what I think, these flaws are produced by coarse statistical >> operations of ANN, & the solution is to start with fine-grain one-to-one >> cross-comparison. >> >>> On Sun, May 10, 2015 at 12:33 PM, Ben Goertzel <[email protected]> wrote: >>> >>> >>>> On Sat, May 9, 2015 at 9:45 PM, Boris Kazachenko <[email protected]> wrote: >>>> Ben, >>>> >>>> a) Episodic memory requires transition from hierarchical to sequential >>>> processing, which is problematic for ANN. Conventional solution is to >>>> model dendrites: they do all the heavy sequential processing, while neuron >>>> itself is more like networking node. Jeff Hawkins does that. >>>> In my model, this dilemma does not exist: every level of a hierarchy is >>>> already a sequence: www.cognitivealgorithm.info >>>> >>>> b) Your solution to pathologies is basically supervision, by some image >>>> grammar. It's a hack, general intelligence must be able to learn without >>>> supervision. I think real solution is in proper node design. >>> >>> >>> No, my suggested solution is NOT supervision by an image grammar. Rather, >>> my preferred approach involves creation of unsupervised learning algorithms >>> that implicitly form image grammar type structures within their networks. >>> Sorry if my write-up was not clear on that point. I agree that most >>> learning in an AGI must occur without supervision, though supervision can >>> also play a supporting role >>> >>> -- Ben >>> AGI | Archives | Modify Your Subscription >> >> AGI | Archives | Modify Your Subscription > > AGI | Archives | Modify Your Subscription ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
