Hello Community, I'm in the last pages of Pedro Domingos' book, the Master Algo, one of two recommended by Bill Gates to learn about AI.
>From the book, I understand all learners have to represent, evaluate, and optimize. There are many types of learners that do this. What Domingos does is generalize these three parts, (1) using Markov Logic Network to represent, (2) posterior probability to evaluate, and (3) genetic search with gradient descent to optimize. The posterior can be replaced for another accuracy measure when it is easier, as genetic search replaced by hill climbing. Where there are 15 popular options for representing, evaluating, and optimizing, Domingos generalized them into three options. The idea is to have one unified learner for any application. There is code already done in R https://alchemy.cs.washington.edu/. My question: anybody in the community vested in coding it into Julia? Thanks. Kevin On Friday, June 3, 2016 at 3:44:09 PM UTC-3, Kevin Liu wrote: > > https://github.com/tbreloff/OnlineAI.jl/issues/5 > > On Friday, June 3, 2016 at 11:17:28 AM UTC-3, Kevin Liu wrote: >> >> I plan to write Julia for the rest of me life... given it remains >> suitable. I am still reading all of Colah's material on nets. I ran >> Mocha.jl a couple weeks ago and was very happy to see it work. Thanks for >> jumping in and telling me about OnlineAI.jl, I will look into it once I am >> ready. From a quick look, perhaps I could help and learn by building a very >> clear documentation of it. Would really like to see Julia a leap ahead of >> other languages, and plan to contribute heavily to it, but at the moment am >> still getting introduced to CS, programming, and nets at the basic level. >> >> On Friday, June 3, 2016 at 10:48:15 AM UTC-3, Tom Breloff wrote: >>> >>> Kevin: computers that program themselves is a concept which is much >>> closer to reality than most would believe, but julia-users isn't really the >>> best place for this speculation. If you're actually interested in writing >>> code, I'm happy to discuss in OnlineAI.jl. I was thinking about how we >>> might tackle code generation using a neural framework I'm working on. >>> >>> On Friday, June 3, 2016, Kevin Liu <[email protected]> wrote: >>> >>>> If Andrew Ng who cited Gates, and Gates who cited Domingos (who did not >>>> lecture at Google with a TensorFlow question in the end), were >>>> unsuccessful >>>> penny traders, Julia was a language for web design, and the tribes in the >>>> video didn't actually solve problems, perhaps this would be a wildly >>>> off-topic, speculative discussion. But these statements couldn't be >>>> farther >>>> from the truth. In fact, if I had known about this video some months ago I >>>> would've understood better on how to solve a problem I was working on. >>>> >>>> For the founders of Julia: I understand your tribe is mainly CS. This >>>> master algorithm, as you are aware, would require collaboration with other >>>> tribes. Just citing the obvious. >>>> >>>> On Friday, June 3, 2016 at 10:21:25 AM UTC-3, Kevin Liu wrote: >>>>> >>>>> There could be parts missing as Domingos mentions, but induction, >>>>> backpropagation, genetic programming, probabilistic inference, and SVMs >>>>> working together-- what's speculative about the improved versions of >>>>> these? >>>>> >>>>> Julia was made for AI. Isn't it time for a consolidated view on how to >>>>> reach it? >>>>> >>>>> On Thursday, June 2, 2016 at 11:20:35 PM UTC-3, Isaiah wrote: >>>>>> >>>>>> This is not a forum for wildly off-topic, speculative discussion. >>>>>> >>>>>> Take this to Reddit, Hacker News, etc. >>>>>> >>>>>> >>>>>> On Thu, Jun 2, 2016 at 10:01 PM, Kevin Liu <[email protected]> wrote: >>>>>> >>>>>>> I am wondering how Julia fits in with the unified tribes >>>>>>> >>>>>>> mashable.com/2016/06/01/bill-gates-ai-code-conference/#8VmBFjIiYOqJ >>>>>>> >>>>>>> https://www.youtube.com/watch?v=B8J4uefCQMc >>>>>>> >>>>>> >>>>>>
