I'll try to build it, compare it, and show it to you guys. I offered to do this as work. I am waiting to see if they will accept it.
On Sunday, August 7, 2016 at 6:15:50 PM UTC-3, Stefan Karpinski wrote: > > Kevin, as previously requested by Isaiah, please take this to some other > forum or maybe start a blog. > > On Sat, Aug 6, 2016 at 10:53 PM, Kevin Liu <kvt...@gmail.com <javascript:> > > wrote: > >> Symmetry-based learning, Domingos, 2014 >> https://www.microsoft.com/en-us/research/video/symmetry-based-learning/ >> >> Approach 2: Deep symmetry networks generalize convolutional neural >> networks by tying parameters and pooling over an arbitrary symmetry group, >> not just the translation group. In preliminary experiments, they >> outperformed convnets on a digit recognition task. >> >> On Friday, August 5, 2016 at 4:56:45 PM UTC-3, Kevin Liu wrote: >>> >>> Minsky died of a cerebral hemorrhage at the age of 88.[40] >>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-40> Ray Kurzweil >>> <https://en.wikipedia.org/wiki/Ray_Kurzweil> says he was contacted by >>> the cryonics organization Alcor Life Extension Foundation >>> <https://en.wikipedia.org/wiki/Alcor_Life_Extension_Foundation> seeking >>> Minsky's body.[41] >>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-Kurzweil-41> >>> Kurzweil >>> believes that Minsky was cryonically preserved by Alcor and will be revived >>> by 2045.[41] >>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-Kurzweil-41> Minsky >>> was a member of Alcor's Scientific Advisory Board >>> <https://en.wikipedia.org/wiki/Advisory_Board>.[42] >>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-AlcorBoard-42> In >>> keeping with their policy of protecting privacy, Alcor will neither confirm >>> nor deny that Alcor has cryonically preserved Minsky.[43] >>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-43> >>> >>> We better do a good job. >>> >>> On Friday, August 5, 2016 at 4:45:42 PM UTC-3, Kevin Liu wrote: >>>> >>>> *So, I think in the next 20 years (2003), if we can get rid of all of >>>> the traditional approaches to artificial intelligence, like neural nets >>>> and >>>> genetic algorithms and rule-based systems, and just turn our sights a >>>> little bit higher to say, can we make a system that can use all those >>>> things for the right kind of problem? Some problems are good for neural >>>> nets; we know that others, neural nets are hopeless on them. Genetic >>>> algorithms are great for certain things; I suspect I know what they're bad >>>> at, and I won't tell you. (Laughter)* - Minsky, founder of CSAIL MIT >>>> >>>> *Those programmers tried to find the single best way to represent >>>> knowledge - Only Logic protects us from paradox.* - Minsky (see >>>> attachment from his lecture) >>>> >>>> On Friday, August 5, 2016 at 8:12:03 AM UTC-3, Kevin Liu wrote: >>>>> >>>>> Markov Logic Network is being used for the continuous development of >>>>> drugs to cure cancer at MIT's CanceRX <http://cancerx.mit.edu/>, on >>>>> DARPA's largest AI project to date, Personalized Assistant that >>>>> Learns (PAL) <https://pal.sri.com/>, progenitor of Siri. One of >>>>> Alchemy's largest applications to date was to learn a semantic network >>>>> (knowledge graph as Google calls it) from the web. >>>>> >>>>> Some on Probabilistic Inductive Logic Programming / Probabilistic >>>>> Logic Programming / Statistical Relational Learning from CSAIL >>>>> <http://people.csail.mit.edu/kersting/ecmlpkdd05_pilp/pilp_ida2005_tut.pdf> >>>>> (my >>>>> understanding is Alchemy does PILP from entailment, proofs, and >>>>> interpretation) >>>>> >>>>> The MIT Probabilistic Computing Project (where there is Picture, an >>>>> extension of Julia, for computer vision; Watch the video from Vikash) >>>>> <http://probcomp.csail.mit.edu/index.html> >>>>> >>>>> Probabilistic programming could do for Bayesian ML what Theano has >>>>> done for neural networks. >>>>> <http://www.inference.vc/deep-learning-is-easy/> - Ferenc Huszár >>>>> >>>>> Picture doesn't appear to be open-source, even though its Paper is >>>>> available. >>>>> >>>>> I'm in the process of comparing the Picture Paper and Alchemy code and >>>>> would like to have an open-source PILP from Julia that combines the best >>>>> of >>>>> both. >>>>> >>>>> On Wednesday, August 3, 2016 at 5:01:02 PM UTC-3, Christof Stocker >>>>> wrote: >>>>>> >>>>>> This sounds like it could be a great contribution. I shall keep a >>>>>> curious eye on your progress >>>>>> >>>>>> Am Mittwoch, 3. August 2016 21:53:54 UTC+2 schrieb Kevin Liu: >>>>>>> >>>>>>> Thanks for the advice Cristof. I am only interested in people >>>>>>> wanting to code it in Julia, from R by Domingos. The algo has been >>>>>>> successfully applied in many areas, even though there are many other >>>>>>> areas >>>>>>> remaining. >>>>>>> >>>>>>> On Wed, Aug 3, 2016 at 4:45 PM, Christof Stocker < >>>>>>> stocker....@gmail.com> wrote: >>>>>>> >>>>>>>> Hello Kevin, >>>>>>>> >>>>>>>> Enthusiasm is a good thing and you should hold on to that. But to >>>>>>>> save yourself some headache or disappointment down the road I advice a >>>>>>>> level head. Nothing is really as bluntly obviously solved as it may >>>>>>>> seems >>>>>>>> at first glance after listening to brilliant people explain things. A >>>>>>>> physics professor of mine once told me that one of the (he thinks) >>>>>>>> most >>>>>>>> malicious factors to his past students progress where overstated >>>>>>>> results/conclusions by other researches (such as premature >>>>>>>> announcements >>>>>>>> from CERN). I am no mathematician, but as far as I can judge is the no >>>>>>>> free >>>>>>>> lunch theorem of pure mathematical nature and not something induced >>>>>>>> empirically. These kind of results are not that easily to get rid of. >>>>>>>> If >>>>>>>> someone (especially an expert) states such a theorem will prove wrong >>>>>>>> I >>>>>>>> would be inclined to believe that he is not talking about literally, >>>>>>>> but >>>>>>>> instead is just trying to make a point about a more or less practical >>>>>>>> implication. >>>>>>>> >>>>>>>> >>>>>>>> Am Mittwoch, 3. August 2016 21:27:05 UTC+2 schrieb Kevin Liu: >>>>>>>>> >>>>>>>>> The Markov logic network represents a probability distribution >>>>>>>>> over the states of a complex system (i.e. a cell), comprised of >>>>>>>>> entities, >>>>>>>>> where logic formulas encode the dependencies between them. >>>>>>>>> >>>>>>>>> On Wednesday, August 3, 2016 at 4:19:09 PM UTC-3, Kevin Liu wrote: >>>>>>>>>> >>>>>>>>>> Alchemy is like an inductive Turing machine, to be programmed to >>>>>>>>>> learn broadly or restrictedly. >>>>>>>>>> >>>>>>>>>> The logic formulas from rules through which it represents can be >>>>>>>>>> inconsistent, incomplete, or even incorrect-- the learning and >>>>>>>>>> probabilistic reasoning will correct them. The key point is that >>>>>>>>>> Alchemy >>>>>>>>>> doesn't have to learn from scratch, proving Wolpert and Macready's >>>>>>>>>> no free >>>>>>>>>> lunch theorem wrong by performing well on a variety of classes of >>>>>>>>>> problems, >>>>>>>>>> not just some. >>>>>>>>>> >>>>>>>>>> On Wednesday, August 3, 2016 at 4:01:15 PM UTC-3, Kevin Liu wrote: >>>>>>>>>>> >>>>>>>>>>> Hello Community, >>>>>>>>>>> >>>>>>>>>>> I'm in the last pages of Pedro Domingos' book, the Master Algo, >>>>>>>>>>> one of two recommended by Bill Gates to learn about AI. >>>>>>>>>>> >>>>>>>>>>> From the book, I understand all learners have to represent, >>>>>>>>>>> evaluate, and optimize. There are many types of learners that do >>>>>>>>>>> this. What >>>>>>>>>>> Domingos does is generalize these three parts, (1) using Markov >>>>>>>>>>> Logic >>>>>>>>>>> Network to represent, (2) posterior probability to evaluate, and >>>>>>>>>>> (3) >>>>>>>>>>> genetic search with gradient descent to optimize. The posterior can >>>>>>>>>>> be >>>>>>>>>>> replaced for another accuracy measure when it is easier, as genetic >>>>>>>>>>> search >>>>>>>>>>> replaced by hill climbing. Where there are 15 popular options for >>>>>>>>>>> representing, evaluating, and optimizing, Domingos generalized them >>>>>>>>>>> into >>>>>>>>>>> three options. The idea is to have one unified learner for any >>>>>>>>>>> application. >>>>>>>>>>> >>>>>>>>>>> There is code already done in R >>>>>>>>>>> https://alchemy.cs.washington.edu/. My question: anybody in the >>>>>>>>>>> community vested in coding it into Julia? >>>>>>>>>>> >>>>>>>>>>> Thanks. Kevin >>>>>>>>>>> >>>>>>>>>>> On Friday, June 3, 2016 at 3:44:09 PM UTC-3, Kevin Liu wrote: >>>>>>>>>>>> >>>>>>>>>>>> https://github.com/tbreloff/OnlineAI.jl/issues/5 >>>>>>>>>>>> >>>>>>>>>>>> On Friday, June 3, 2016 at 11:17:28 AM UTC-3, Kevin Liu wrote: >>>>>>>>>>>>> >>>>>>>>>>>>> I plan to write Julia for the rest of me life... given it >>>>>>>>>>>>> remains suitable. I am still reading all of Colah's material on >>>>>>>>>>>>> nets. I ran >>>>>>>>>>>>> Mocha.jl a couple weeks ago and was very happy to see it work. >>>>>>>>>>>>> Thanks for >>>>>>>>>>>>> jumping in and telling me about OnlineAI.jl, I will look into it >>>>>>>>>>>>> once I am >>>>>>>>>>>>> ready. From a quick look, perhaps I could help and learn by >>>>>>>>>>>>> building a very >>>>>>>>>>>>> clear documentation of it. Would really like to see Julia a leap >>>>>>>>>>>>> ahead of >>>>>>>>>>>>> other languages, and plan to contribute heavily to it, but at the >>>>>>>>>>>>> moment am >>>>>>>>>>>>> still getting introduced to CS, programming, and nets at the >>>>>>>>>>>>> basic level. >>>>>>>>>>>>> >>>>>>>>>>>>> On Friday, June 3, 2016 at 10:48:15 AM UTC-3, Tom Breloff >>>>>>>>>>>>> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>> Kevin: computers that program themselves is a concept which >>>>>>>>>>>>>> is much closer to reality than most would believe, but >>>>>>>>>>>>>> julia-users isn't >>>>>>>>>>>>>> really the best place for this speculation. If you're actually >>>>>>>>>>>>>> interested >>>>>>>>>>>>>> in writing code, I'm happy to discuss in OnlineAI.jl. I was >>>>>>>>>>>>>> thinking about >>>>>>>>>>>>>> how we might tackle code generation using a neural framework I'm >>>>>>>>>>>>>> working >>>>>>>>>>>>>> on. >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Friday, June 3, 2016, Kevin Liu <kvt...@gmail.com> wrote: >>>>>>>>>>>>>> >>>>>>>>>>>>>>> If Andrew Ng who cited Gates, and Gates who cited Domingos >>>>>>>>>>>>>>> (who did not lecture at Google with a TensorFlow question in >>>>>>>>>>>>>>> the end), were >>>>>>>>>>>>>>> unsuccessful penny traders, Julia was a language for web >>>>>>>>>>>>>>> design, and the >>>>>>>>>>>>>>> tribes in the video didn't actually solve problems, perhaps >>>>>>>>>>>>>>> this would be a >>>>>>>>>>>>>>> wildly off-topic, speculative discussion. But these statements >>>>>>>>>>>>>>> couldn't be >>>>>>>>>>>>>>> farther from the truth. In fact, if I had known about this >>>>>>>>>>>>>>> video some >>>>>>>>>>>>>>> months ago I would've understood better on how to solve a >>>>>>>>>>>>>>> problem I was >>>>>>>>>>>>>>> working on. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> For the founders of Julia: I understand your tribe is mainly >>>>>>>>>>>>>>> CS. This master algorithm, as you are aware, would require >>>>>>>>>>>>>>> collaboration >>>>>>>>>>>>>>> with other tribes. Just citing the obvious. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Friday, June 3, 2016 at 10:21:25 AM UTC-3, Kevin Liu >>>>>>>>>>>>>>> wrote: >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> There could be parts missing as Domingos mentions, but >>>>>>>>>>>>>>>> induction, backpropagation, genetic programming, probabilistic >>>>>>>>>>>>>>>> inference, >>>>>>>>>>>>>>>> and SVMs working together-- what's speculative about the >>>>>>>>>>>>>>>> improved versions >>>>>>>>>>>>>>>> of these? >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Julia was made for AI. Isn't it time for a consolidated >>>>>>>>>>>>>>>> view on how to reach it? >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> On Thursday, June 2, 2016 at 11:20:35 PM UTC-3, Isaiah >>>>>>>>>>>>>>>> wrote: >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> This is not a forum for wildly off-topic, speculative >>>>>>>>>>>>>>>>> discussion. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> Take this to Reddit, Hacker News, etc. >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> On Thu, Jun 2, 2016 at 10:01 PM, Kevin Liu < >>>>>>>>>>>>>>>>> kvt...@gmail.com> wrote: >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> I am wondering how Julia fits in with the unified tribes >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> mashable.com/2016/06/01/bill-gates-ai-code-conference/#8VmBFjIiYOqJ >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> https://www.youtube.com/watch?v=B8J4uefCQMc >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> >>>>>>> >