These days a symbolic system is usually seen in the form of a network - as almost everyone in this groups know. The idea that a symbolic network will need deep NNs is seems like it is a little obscure except as an immediate practical matter. Jim Bromer
On Sun, Feb 17, 2019 at 8:27 AM Ben Goertzel <b...@goertzel.org> wrote: > One can see the next steps from the analogy of deep NNs for computer > vision > > First they did straightforward visual analytics, then they started > worrying more about the internal representations, and now in the last > 6 months or so there is finally a little progress in getting sensible > internal representations within deep NNs analyzing visual scenes. > > Don't get me wrong tho, I don't think this is the golden path to AGI > or anything.... However, the next step is clearly to try to tweak the > architecture to get more transparent internal representations. As it > happens this would also be useful for interfacing such deep NNs with > symbolic systems or other sorts of AI algorithms... > > -- Ben > > On Sun, Feb 17, 2019 at 9:05 PM Stefan Reich via AGI > <agi@agi.topicbox.com> wrote: > > > > I'm not sure how one would go the next step from a > random-speech-generating network like that. > > > > We do want the speech to mean something. > > > > My new approach is to incorporate semantics into a rule engine right > from the start. > > > > On Sun, 17 Feb 2019 at 02:09, Ben Goertzel <b...@goertzel.org> wrote: > >> > >> Rob, > >> > >> These deep NNs certainly are not linear models, and they do capture a > >> bunch of syntactic phenomena fairly subtly, see e.g. > >> > >> https://arxiv.org/abs/1901.05287 > >> > >> "I assess the extent to which the recently introduced BERT model > >> captures English syntactic phenomena, using (1) naturally-occurring > >> subject-verb agreement stimuli; (2) "coloreless green ideas" > >> subject-verb agreement stimuli, in which content words in natural > >> sentences are randomly replaced with words sharing the same > >> part-of-speech and inflection; and (3) manually crafted stimuli for > >> subject-verb agreement and reflexive anaphora phenomena. The BERT > >> model performs remarkably well on all cases." > >> > >> This paper shows some dependency trees implicit in transformer networks, > >> > >> http://aclweb.org/anthology/W18-5431 > >> > >> This stuff is not AGI and does not extract deep semantics nor do > >> symbol grounding etc. For sure it has many limitations. Bu it's > >> also not so trivial as you're suggesting IMO... > >> > >> -- Ben G > >> > >> On Sun, Feb 17, 2019 at 8:42 AM Rob Freeman <chaotic.langu...@gmail.com> > wrote: > >> > > >> > On the substance, here's what I wrote elsewhere in response to > someone's comment that it is an "important step": > >> > > >> > Important step? I don't see it. Bengio's NLM? Yeah, good, we need > distributed representation. That was an advance. but it was always a linear > model without a sensible way of folding in context. Now they try to fold in > a bit of context by bolting on another layer to spotlight other parts of > the sequence ad-hoc? > >> > > >> > I don't see any theoretical cohesiveness, any actual theory let alone > novelty of theory. > >> > > >> > What is the underlying model for language here? In particular what is > the underlying model for how words combine to create meaning? How do parts > of a sequence combine to become a whole, incorporating the whole context? > Linear combination with a bolt-on spotlight? > >> > > >> > I think all this ad-hoc tinkering will be thrown away when we figure > out a principled way to combine words which incorporates context > inherently. But nobody is even attempting that. They are just tinkering. > Limited to tinkering with linear models, because nothing else can be > "learned". > >> > > >> > On Sun, Feb 17, 2019 at 1:05 PM Ben Goertzel <b...@goertzel.org> > wrote: > >> >> > >> >> Hmmm... > >> >> > >> >> About this "OpenAI keeping their language model secret" thing... > >> >> > >> >> I mean -- clearly, keeping their language model secret is a pure PR > >> >> stunt... Their > >> >> algorithm is described in an online paper... and their model was > >> >> trained on Reddit text ... so anyone else with a bunch of $$ (for > >> >> machine-time and data-preprocessing hacking) can download Reddit > >> >> (complete Reddit archives are available as a torrent) and train a > >> >> language model similar or better > >> >> than OpenAI's ... > >> >> > >> >> That said, their language model is a moderate improvement on the BERT > >> >> model released by Google last year. This is good AI work. There is > >> >> no understanding of semantics and no grounding of symbols in > >> >> experience/world here, but still, it's pretty f**king cool to see > what > >> >> an awesome job of text generation can be done by these pure > >> >> surface-level-pattern-recognition methods.... > >> >> > >> >> Honestly a lot of folks in the deep-NN/NLP space (including our own > >> >> SingularityNET St. Petersburg team) have been talking about applying > >> >> BERT-ish attention networks (with more comprehensive network > >> >> architectures) in similar ways... but there are always so many > >> >> different things to work on, and OpenAI should be congratulated for > >> >> making these particular architecture tweaks and demonstrating them > >> >> first... but not for the PR stunt of keeping their model secret... > >> >> > >> >> Although perhaps they should be congratulated for revealing so > clearly > >> >> the limitations of the "open-ness" in their name "Open AI." I mean, > >> >> we all know there are some cases where keeping something secret may > be > >> >> the most ethical choice ... but the fact that they're willing to take > >> >> this step simply for a short-term one-news-cycle PR boost, indicates > >> >> that open-ness may not be such an important value to them after > all... > >> >> > >> >> -- > >> >> Ben Goertzel, PhD > >> >> http://goertzel.org > >> >> > >> >> "Listen: This world is the lunatic's sphere, / Don't always agree > >> >> it's real. / Even with my feet upon it / And the postman knowing my > >> >> door / My address is somewhere else." -- Hafiz > >> > > >> > Artificial General Intelligence List / AGI / see discussions + > participants + delivery options Permalink > >> > >> > >> -- > >> Ben Goertzel, PhD > >> http://goertzel.org > >> > >> "Listen: This world is the lunatic's sphere, / Don't always agree > >> it's real. / Even with my feet upon it / And the postman knowing my > >> door / My address is somewhere else." -- Hafiz > > > > > > > > -- > > Stefan Reich > > BotCompany.de // Java-based operating systems > > Artificial General Intelligence List / AGI / see discussions + > participants + delivery options Permalink > > -- > Ben Goertzel, PhD > http://goertzel.org > > "Listen: This world is the lunatic's sphere, / Don't always agree > it's real. / Even with my feet upon it / And the postman knowing my > door / My address is somewhere else." -- Hafiz ------------------------------------------ Artificial General Intelligence List: AGI Permalink: https://agi.topicbox.com/groups/agi/T581199cf280badd7-M3b0eb7cb8ec10f83acc920bb Delivery options: https://agi.topicbox.com/groups/agi/subscription