> As I think about it, one problem is, depending on how its > parametrized, its not going to build much of a world model. > Say for example it uses trigrams. The average hs grad knows > something like 50,000 words. So there are something like 10^17 > trigrams. It will never see enough data to build a model capturing > much semantics, unless it builds an incredibly compact model, > in which case-- what is the underlying structure and how > (computationally) are you going to learn it?
Absolutely correct. That's why I said "My belief is that if you had the proper structure-building learning algorithms that your operator grammar system would simply (re-)discover the basic parts of speech and would then successfully proceed from there." and why I slammed it for ""reinventing the wheel" in terms of it's unnecessary generalization of dependency" > In unsupervised learning, you can learn a lot, > say you can cluster the world into two clusters. But until you get > supervision, you can't learn the final few bits to distinguish good > from bad, or whatever. I'm afraid that I disagree completely with the latter sentence. > Operator grammar might be very useful for > getting a structure that could then be rapidly trained to produce > meaning, but I don't think you can finish the job until you interact > with sensation. It seems as if you're now talking sensory fusion (which is a whole 'nother can o' worms). Mark ----- This list is sponsored by AGIRI: http://www.agiri.org/email To unsubscribe or change your options, please go to: http://v2.listbox.com/member/?member_id=231415&user_secret=e9e40a7e