Thanks for the advice Cristof. I am only interested in people wanting to
code it in Julia, from R by Domingos. The algo has been successfully
applied in many areas, even though there are many other areas remaining.

On Wed, Aug 3, 2016 at 4:45 PM, Christof Stocker <[email protected]
> wrote:

> Hello Kevin,
>
> Enthusiasm is a good thing and you should hold on to that. But to save
> yourself some headache or disappointment down the road I advice a level
> head. Nothing is really as bluntly obviously solved as it may seems at
> first glance after listening to brilliant people explain things. A physics
> professor of mine once told me that one of the (he thinks) most malicious
> factors to his past students progress where overstated results/conclusions
> by other researches (such as premature announcements from CERN). I am no
> mathematician, but as far as I can judge is the no free lunch theorem of
> pure mathematical nature and not something induced empirically. These kind
> of results are not that easily to get rid of. If someone (especially an
> expert) states such a theorem will prove wrong I would be inclined to
> believe that he is not talking about literally, but instead is just trying
> to make a point about a more or less practical implication.
>
>
> Am Mittwoch, 3. August 2016 21:27:05 UTC+2 schrieb Kevin Liu:
>>
>> The Markov logic network represents a probability distribution over the
>> states of a complex system (i.e. a cell), comprised of entities, where
>> logic formulas encode the dependencies between them.
>>
>> On Wednesday, August 3, 2016 at 4:19:09 PM UTC-3, Kevin Liu wrote:
>>>
>>> Alchemy is like an inductive Turing machine, to be programmed to learn
>>> broadly or restrictedly.
>>>
>>> The logic formulas from rules through which it represents can be
>>> inconsistent, incomplete, or even incorrect-- the learning and
>>> probabilistic reasoning will correct them. The key point is that Alchemy
>>> doesn't have to learn from scratch, proving Wolpert and Macready's no free
>>> lunch theorem wrong by performing well on a variety of classes of problems,
>>> not just some.
>>>
>>> On Wednesday, August 3, 2016 at 4:01:15 PM UTC-3, Kevin Liu wrote:
>>>>
>>>> Hello Community,
>>>>
>>>> I'm in the last pages of Pedro Domingos' book, the Master Algo, one of
>>>> two recommended by Bill Gates to learn about AI.
>>>>
>>>> From the book, I understand all learners have to represent, evaluate,
>>>> and optimize. There are many types of learners that do this. What Domingos
>>>> does is generalize these three parts, (1) using Markov Logic Network to
>>>> represent, (2) posterior probability to evaluate, and (3) genetic search
>>>> with gradient descent to optimize. The posterior can be replaced for
>>>> another accuracy measure when it is easier, as genetic search replaced by
>>>> hill climbing. Where there are 15 popular options for representing,
>>>> evaluating, and optimizing, Domingos generalized them into three options.
>>>> The idea is to have one unified learner for any application.
>>>>
>>>> There is code already done in R https://alchemy.cs.washington.edu/. My
>>>> question: anybody in the community vested in coding it into Julia?
>>>>
>>>> Thanks. Kevin
>>>>
>>>> On Friday, June 3, 2016 at 3:44:09 PM UTC-3, Kevin Liu wrote:
>>>>>
>>>>> https://github.com/tbreloff/OnlineAI.jl/issues/5
>>>>>
>>>>> On Friday, June 3, 2016 at 11:17:28 AM UTC-3, Kevin Liu wrote:
>>>>>>
>>>>>> I plan to write Julia for the rest of me life... given it remains
>>>>>> suitable. I am still reading all of Colah's material on nets. I ran
>>>>>> Mocha.jl a couple weeks ago and was very happy to see it work. Thanks for
>>>>>> jumping in and telling me about OnlineAI.jl, I will look into it once I 
>>>>>> am
>>>>>> ready. From a quick look, perhaps I could help and learn by building a 
>>>>>> very
>>>>>> clear documentation of it. Would really like to see Julia a leap ahead of
>>>>>> other languages, and plan to contribute heavily to it, but at the moment 
>>>>>> am
>>>>>> still getting introduced to CS, programming, and nets at the basic level.
>>>>>>
>>>>>> On Friday, June 3, 2016 at 10:48:15 AM UTC-3, Tom Breloff wrote:
>>>>>>>
>>>>>>> Kevin: computers that program themselves is a concept which is much
>>>>>>> closer to reality than most would believe, but julia-users isn't really 
>>>>>>> the
>>>>>>> best place for this speculation. If you're actually interested in 
>>>>>>> writing
>>>>>>> code, I'm happy to discuss in OnlineAI.jl. I was thinking about how we
>>>>>>> might tackle code generation using a neural framework I'm working on.
>>>>>>>
>>>>>>> On Friday, June 3, 2016, Kevin Liu <[email protected]> wrote:
>>>>>>>
>>>>>>>> If Andrew Ng who cited Gates, and Gates who cited Domingos (who did
>>>>>>>> not lecture at Google with a TensorFlow question in the end), were
>>>>>>>> unsuccessful penny traders, Julia was a language for web design, and 
>>>>>>>> the
>>>>>>>> tribes in the video didn't actually solve problems, perhaps this would 
>>>>>>>> be a
>>>>>>>> wildly off-topic, speculative discussion. But these statements 
>>>>>>>> couldn't be
>>>>>>>> farther from the truth. In fact, if I had known about this video some
>>>>>>>> months ago I would've understood better on how to solve a problem I was
>>>>>>>> working on.
>>>>>>>>
>>>>>>>> For the founders of Julia: I understand your tribe is mainly CS.
>>>>>>>> This master algorithm, as you are aware, would require collaboration 
>>>>>>>> with
>>>>>>>> other tribes. Just citing the obvious.
>>>>>>>>
>>>>>>>> On Friday, June 3, 2016 at 10:21:25 AM UTC-3, Kevin Liu wrote:
>>>>>>>>>
>>>>>>>>> There could be parts missing as Domingos mentions, but induction,
>>>>>>>>> backpropagation, genetic programming, probabilistic inference, and 
>>>>>>>>> SVMs
>>>>>>>>> working together-- what's speculative about the improved versions of 
>>>>>>>>> these?
>>>>>>>>>
>>>>>>>>> Julia was made for AI. Isn't it time for a consolidated view on
>>>>>>>>> how to reach it?
>>>>>>>>>
>>>>>>>>> On Thursday, June 2, 2016 at 11:20:35 PM UTC-3, Isaiah wrote:
>>>>>>>>>>
>>>>>>>>>> This is not a forum for wildly off-topic, speculative discussion.
>>>>>>>>>>
>>>>>>>>>> Take this to Reddit, Hacker News, etc.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Thu, Jun 2, 2016 at 10:01 PM, Kevin Liu <[email protected]>
>>>>>>>>>> wrote:
>>>>>>>>>>
>>>>>>>>>>> I am wondering how Julia fits in with the unified tribes
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> mashable.com/2016/06/01/bill-gates-ai-code-conference/#8VmBFjIiYOqJ
>>>>>>>>>>>
>>>>>>>>>>> https://www.youtube.com/watch?v=B8J4uefCQMc
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>

Reply via email to