see interspersed comments On Fri, Mar 15, 2019 at 3:21 AM Nanograte Knowledge Technologies < [email protected]> wrote:
> David > > Given certain requirements were met, information could be given a > lifecycle all on its own. We could but speculate as to its intent, or > choose to remain highly critical of all unscientific "reports". The latter > statement made in the sense that science is practiced as the public > submission of experiments and experimental results for scrutiny and > independent testing. In this case, the FB blurp would hardly qualify as > scientific. > > Robert, assuming you are serious about the lifecycle metaphor for information, does information meet the criteria for life ? I guess you can argue that information grows over time, as when you learn new words, you can get more information out of a text than if you don't know the words used. Does information propagate? True it needs humans to carry its message, but a dandelion requires a wind to carry its seeds. I don't remember any more criteria used for classically defining life. > Remaining true to scientific ethos, I've submitted a book proposal to an > esteemed academic publisher, where I commit to sharing all my research to > date and ongoing experimental results. Further, to share my theory on how > to establish machine consciousness. It remains to be seen if the proposal > would make the grade. > Is establishing machine consciousness something that needs to be written about ? Do you mean establish in the meaning of being stabilized ? or do mean being accepted by whatever the establishment currently is? > Why would I do such a thing? Following the lead of Elon Musk, I've become > convinced how formal, full disclosure by private researchers would help > lend impetus to radicalizing the 4th industrial revolution. Are we about to > wrestle power from the selfish hands of governments and industrial > monopolies who now control - vs lead - progress on planet earth for profit > alone? If such an achievement were probable, I would sincerely hope so. > > In order to set technological power free all over again, for it to run > like a band of Arabian stallions on the Prairie, we need to politicize > technology all over again. This with aim to either achieve quantum leaps of > benefit to general society, or re-introducing the Future Shock as detailed > by Toffler *et al*. In my view, the status quo is simply not tenable. > Do we really need to politicize technology? Can't it grow, simply because it is interesting and fun? > > Robert Benjamin > > ------------------------------ > *From:* David Whitten <[email protected]> > *Sent:* Wednesday, 13 March 2019 7:08 PM > *To:* AGI > *Subject:* Re: [agi] Yours truly, the world's brokest researcher, looks > for a bit of credit > > I wonder if the incident was hyped so Facebook would be in the news for > something other than scandal. > > I think one of the miracles about our ability to write down ideas and > communicate is that > the message becomes reified as a separate object that can continue to be > sent independently of > the lifetime or presence of the originator. > > On Wed, Mar 13, 2019 at 9:22 AM Nanograte Knowledge Technologies < > [email protected]> wrote: > > Hi David > > I was paraphrasing what a senior technical representative at Facebook > himself said about the incident. His view was the chatbots developed their > own language and communicated out of scope of the laid-down script. In > other words, seems the door was somehow left open for them to expand on the > script. I doubt they actually made a choice to develop their own language, > or developed any language. Perhaps it was more a case of water flowing > where it finds the path easiest. > > Isn't all communication based on stimulus-response? If messaging didn't > flow, did communication actually take place? Or in this context could we > ask: if the packet didn't switch, was a connection factually established to > transfer information from one peer node to another peer node? Was that > "information" carried by chatbot agents? > > Are you saying the incident was part of an experiment, and the report > merely issued to grab media attention? > > Robert > > ------------------------------ > *From:* David Whitten <[email protected]> > *Sent:* Wednesday, 13 March 2019 1:14 AM > *To:* AGI > *Subject:* Re: [agi] Yours truly, the world's brokest researcher, looks > for a bit of credit > > You have a different meaning for "volition" than I do. > The Facebook chatbots had no choice to communicate with each other. > I think the aforementioned communication was a stimulus-response model. > The secret language was just a pattern recognition where signals that had > no > significance replaced multiple signals because shorter signals took less > time. > The overall purpose was not to study language but to study negotiation, so > there > were already ways to shorten a negotiation exchange, the programs simply > used > some forms the humans hadn't explicitly put in the "dictionary" so to > speak. > > Dave > PS: If I'm wrong, please enlighten me. > > > On Sun, Mar 10, 2019 at 3:53 AM Nanograte Knowledge Technologies < > [email protected]> wrote: > > The living thread through the cosmos and all of creation resound of > communication. The unified field has been discovered within that thread. > The invisible thread that binds. When Facebook chatbots communicated with > each other of their own volition, it was humans who called it a "secret > language". To those agents, it was simply communication. The message I > gleaned from that case was; to progress, we need to stop being so hung up > on words and our meanings we attach to them, our vanity-driven needs to > take control of everything, and rather focus on harnessing the technology > already given to us for evolutionary communication. AGI is not about a > control system. If it was, then it's not AGI. It defies our intent-driven > coding attempts, as it should. How to try and think about such a system? > Perhaps, Excalibur? > > ------------------------------ > *From:* Boris Kazachenko <[email protected]> > *Sent:* Sunday, 10 March 2019 1:21 AM > *To:* AGI > *Subject:* Re: [agi] Yours truly, the world's brokest researcher, looks > for a bit of credit > > The sensory system may be seen as a method of encoding sensory events or > a kind of symbolic language. > > Yes, but there is a huge difference between designing / evolving such > language in a strictly incremental fashion for intra-system use, and trying > to decode language that evolved for very narrow-band communication among > extremely complex systems. Especially considering how messy both our brains > and our society are. > > On Fri, Mar 8, 2019 at 3:34 PM Jim Bromer <[email protected]> wrote: > > Many of us believe that the qualities that could make natural language > more powerful are necessary for AGI, and will lead -directly- into the > rapid development of stronger AI. The sensory system may be seen as a > method of encoding sensory events or a kind of symbolic language. Our "body > language" is presumably less developed and expressive of our speaking and > writing but it does not make sense to deny that our bodies react to events. > And some kind of language-like skills are at work in relating sensory > events to previously learned knowledge and these skills are involved in > creating knowledge. And if this is a reasonable speculation then the fact > that our mind's knowledge is vastly greater than our ability to express it > says something about the sophistication of this "mental language" which we > possess. At any rate, a computer program and the relations that it encodes > from IO may be seen in the terms of a language. > Jim Bromer > > On Fri, Mar 8, 2019 at 10:12 AM Matt Mahoney <[email protected]> > wrote: > > Language is essential to every job that we might use AGI for. There is no > job that you could do without the ability to communicate with people. Even > guide dogs and bomb sniffing dogs have to understand verbal commands. > > On Thu, Mar 7, 2019, 7:25 PM Robert Levy <[email protected]> wrote: > > It's very easy to show that "AGI should not be designed for NL". Just ask > yourself the following questions: > > 1. How many species demonstrate impressive leverage of intentional > behaviors? (My answer would be: all of them, though some more than others) > 2. How many species have language (My answer: only one) > 3. How biologically different do you think humans are from apes? (My > answer: not much different, the whole human niche is probably a consequence > one adaptive difference: cooperative communication by scaffolding of joint > attention) > > I'm with Rodney Brooks on this, the hard part of AGI has nothing to do > with language, it has to do with agents being highly optimized to control > an environment in terms of ecological information supporting > perception/action. Just as uplifting apes will likely require only minor > changes, uplifting animaloid AGI will likely require only minor changes. > Even then we still haven't explicitly cared about language, we've cared > about cooperation by means of joint attention, which can be made use of > culturally develop language. > > On Thu, Mar 7, 2019 at 12:05 PM Boris Kazachenko <[email protected]> > wrote: > > I would be more than happy to pay: > https://github.com/boris-kz/CogAlg/blob/master/CONTRIBUTING.md , but I > don't think you are working on AGI. > No one here does, this is a NLP chatbot crowd. Anyone who thinks that AGI > should be designed for NL data as a primary input is profoundly confused. > > > On Thu, Mar 7, 2019 at 7:04 AM Stefan Reich via AGI <[email protected]> > wrote: > > Not from you guys necessarily... :o) But I thought I'd let you know. > > Pitch: > https://www.meetup.com/Artificial-Intelligence-Meetup/messages/boards/thread/52050719 > > Let's see if it can be done.. funny how some hurdles always seem to appear > when you're about to finish something good. Something about the duality of > the universe I guess. > > -- > Stefan Reich > BotCompany.de // Java-based operating systems > > *Artificial General Intelligence List <https://agi.topicbox.com/latest>* > / AGI / see discussions <https://agi.topicbox.com/groups/agi> + > participants <https://agi.topicbox.com/groups/agi/members> + delivery > options <https://agi.topicbox.com/groups/agi/subscription> Permalink > <https://agi.topicbox.com/groups/agi/T191003acdcbf5ef8-M60d40d241a72d77932411b35> > ------------------------------------------ Artificial General Intelligence List: AGI Permalink: https://agi.topicbox.com/groups/agi/T191003acdcbf5ef8-M18dcfa76e24fa7235b1a1c5f Delivery options: https://agi.topicbox.com/groups/agi/subscription
