After which I have to code Felix into Julia, a relational optimizer for 
statistical inference with Tuffy <http://i.stanford.edu/hazy/tuffy/> 
inside, for enterprise settings.

On Tuesday, August 9, 2016 at 12:07:32 AM UTC-3, Kevin Liu wrote:
>
> Can I get tips on bringing Alchemy's optimized Tuffy 
> <http://i.stanford.edu/hazy/tuffy/> in Java to Julia while showing the 
> best of Julia? I am going for the most correct way, even if it means coding 
> Tuffy into C and Julia.
>
> On Sunday, August 7, 2016 at 8:34:37 PM UTC-3, Kevin Liu wrote:
>>
>> I'll try to build it, compare it, and show it to you guys. I offered to 
>> do this as work. I am waiting to see if they will accept it. 
>>
>> On Sunday, August 7, 2016 at 6:15:50 PM UTC-3, Stefan Karpinski wrote:
>>>
>>> Kevin, as previously requested by Isaiah, please take this to some other 
>>> forum or maybe start a blog.
>>>
>>> On Sat, Aug 6, 2016 at 10:53 PM, Kevin Liu <kvt...@gmail.com> wrote:
>>>
>>>> Symmetry-based learning, Domingos, 2014 
>>>> https://www.microsoft.com/en-us/research/video/symmetry-based-learning/
>>>>
>>>> Approach 2: Deep symmetry networks generalize convolutional neural 
>>>> networks by tying parameters and pooling over an arbitrary symmetry group, 
>>>> not just the translation group. In preliminary experiments, they 
>>>> outperformed convnets on a digit recognition task. 
>>>>
>>>> On Friday, August 5, 2016 at 4:56:45 PM UTC-3, Kevin Liu wrote:
>>>>>
>>>>> Minsky died of a cerebral hemorrhage at the age of 88.[40] 
>>>>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-40> Ray 
>>>>> Kurzweil <https://en.wikipedia.org/wiki/Ray_Kurzweil> says he was 
>>>>> contacted by the cryonics organization Alcor Life Extension Foundation 
>>>>> <https://en.wikipedia.org/wiki/Alcor_Life_Extension_Foundation> seeking 
>>>>> Minsky's body.[41] 
>>>>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-Kurzweil-41> 
>>>>> Kurzweil 
>>>>> believes that Minsky was cryonically preserved by Alcor and will be 
>>>>> revived 
>>>>> by 2045.[41] 
>>>>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-Kurzweil-41> 
>>>>> Minsky 
>>>>> was a member of Alcor's Scientific Advisory Board 
>>>>> <https://en.wikipedia.org/wiki/Advisory_Board>.[42] 
>>>>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-AlcorBoard-42> In 
>>>>> keeping with their policy of protecting privacy, Alcor will neither 
>>>>> confirm 
>>>>> nor deny that Alcor has cryonically preserved Minsky.[43] 
>>>>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-43> 
>>>>>
>>>>> We better do a good job. 
>>>>>
>>>>> On Friday, August 5, 2016 at 4:45:42 PM UTC-3, Kevin Liu wrote:
>>>>>>
>>>>>> *So, I think in the next 20 years (2003), if we can get rid of all of 
>>>>>> the traditional approaches to artificial intelligence, like neural nets 
>>>>>> and 
>>>>>> genetic algorithms and rule-based systems, and just turn our sights a 
>>>>>> little bit higher to say, can we make a system that can use all those 
>>>>>> things for the right kind of problem? Some problems are good for neural 
>>>>>> nets; we know that others, neural nets are hopeless on them. Genetic 
>>>>>> algorithms are great for certain things; I suspect I know what they're 
>>>>>> bad 
>>>>>> at, and I won't tell you. (Laughter)*  - Minsky, founder of CSAIL MIT
>>>>>>
>>>>>> *Those programmers tried to find the single best way to represent 
>>>>>> knowledge - Only Logic protects us from paradox.* - Minsky (see 
>>>>>> attachment from his lecture)
>>>>>>
>>>>>> On Friday, August 5, 2016 at 8:12:03 AM UTC-3, Kevin Liu wrote:
>>>>>>>
>>>>>>> Markov Logic Network is being used for the continuous development of 
>>>>>>> drugs to cure cancer at MIT's CanceRX <http://cancerx.mit.edu/>, on 
>>>>>>> DARPA's largest AI project to date, Personalized Assistant that 
>>>>>>> Learns (PAL) <https://pal.sri.com/>, progenitor of Siri. One of 
>>>>>>> Alchemy's largest applications to date was to learn a semantic network 
>>>>>>> (knowledge graph as Google calls it) from the web. 
>>>>>>>
>>>>>>> Some on Probabilistic Inductive Logic Programming / Probabilistic 
>>>>>>> Logic Programming / Statistical Relational Learning from CSAIL 
>>>>>>> <http://people.csail.mit.edu/kersting/ecmlpkdd05_pilp/pilp_ida2005_tut.pdf>
>>>>>>>  (my 
>>>>>>> understanding is Alchemy does PILP from entailment, proofs, and 
>>>>>>> interpretation)
>>>>>>>
>>>>>>> The MIT Probabilistic Computing Project (where there is Picture, an 
>>>>>>> extension of Julia, for computer vision; Watch the video from Vikash) 
>>>>>>> <http://probcomp.csail.mit.edu/index.html>
>>>>>>>
>>>>>>> Probabilistic programming could do for Bayesian ML what Theano has 
>>>>>>> done for neural networks. 
>>>>>>> <http://www.inference.vc/deep-learning-is-easy/> - Ferenc Huszár
>>>>>>>
>>>>>>> Picture doesn't appear to be open-source, even though its Paper is 
>>>>>>> available. 
>>>>>>>
>>>>>>> I'm in the process of comparing the Picture Paper and Alchemy code 
>>>>>>> and would like to have an open-source PILP from Julia that combines the 
>>>>>>> best of both. 
>>>>>>>
>>>>>>> On Wednesday, August 3, 2016 at 5:01:02 PM UTC-3, Christof Stocker 
>>>>>>> wrote:
>>>>>>>>
>>>>>>>> This sounds like it could be a great contribution. I shall keep a 
>>>>>>>> curious eye on your progress
>>>>>>>>
>>>>>>>> Am Mittwoch, 3. August 2016 21:53:54 UTC+2 schrieb Kevin Liu:
>>>>>>>>>
>>>>>>>>> Thanks for the advice Cristof. I am only interested in people 
>>>>>>>>> wanting to code it in Julia, from R by Domingos. The algo has been 
>>>>>>>>> successfully applied in many areas, even though there are many other 
>>>>>>>>> areas 
>>>>>>>>> remaining. 
>>>>>>>>>
>>>>>>>>> On Wed, Aug 3, 2016 at 4:45 PM, Christof Stocker <
>>>>>>>>> stocker....@gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> Hello Kevin,
>>>>>>>>>>
>>>>>>>>>> Enthusiasm is a good thing and you should hold on to that. But to 
>>>>>>>>>> save yourself some headache or disappointment down the road I advice 
>>>>>>>>>> a 
>>>>>>>>>> level head. Nothing is really as bluntly obviously solved as it may 
>>>>>>>>>> seems 
>>>>>>>>>> at first glance after listening to brilliant people explain things. 
>>>>>>>>>> A 
>>>>>>>>>> physics professor of mine once told me that one of the (he thinks) 
>>>>>>>>>> most 
>>>>>>>>>> malicious factors to his past students progress where overstated 
>>>>>>>>>> results/conclusions by other researches (such as premature 
>>>>>>>>>> announcements 
>>>>>>>>>> from CERN). I am no mathematician, but as far as I can judge is the 
>>>>>>>>>> no free 
>>>>>>>>>> lunch theorem of pure mathematical nature and not something induced 
>>>>>>>>>> empirically. These kind of results are not that easily to get rid 
>>>>>>>>>> of. If 
>>>>>>>>>> someone (especially an expert) states such a theorem will prove 
>>>>>>>>>> wrong I 
>>>>>>>>>> would be inclined to believe that he is not talking about literally, 
>>>>>>>>>> but 
>>>>>>>>>> instead is just trying to make a point about a more or less 
>>>>>>>>>> practical 
>>>>>>>>>> implication.
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> Am Mittwoch, 3. August 2016 21:27:05 UTC+2 schrieb Kevin Liu:
>>>>>>>>>>>
>>>>>>>>>>> The Markov logic network represents a probability distribution 
>>>>>>>>>>> over the states of a complex system (i.e. a cell), comprised of 
>>>>>>>>>>> entities, 
>>>>>>>>>>> where logic formulas encode the dependencies between them. 
>>>>>>>>>>>
>>>>>>>>>>> On Wednesday, August 3, 2016 at 4:19:09 PM UTC-3, Kevin Liu 
>>>>>>>>>>> wrote:
>>>>>>>>>>>>
>>>>>>>>>>>> Alchemy is like an inductive Turing machine, to be programmed 
>>>>>>>>>>>> to learn broadly or restrictedly.
>>>>>>>>>>>>
>>>>>>>>>>>> The logic formulas from rules through which it represents can 
>>>>>>>>>>>> be inconsistent, incomplete, or even incorrect-- the learning and 
>>>>>>>>>>>> probabilistic reasoning will correct them. The key point is that 
>>>>>>>>>>>> Alchemy 
>>>>>>>>>>>> doesn't have to learn from scratch, proving Wolpert and Macready's 
>>>>>>>>>>>> no free 
>>>>>>>>>>>> lunch theorem wrong by performing well on a variety of classes of 
>>>>>>>>>>>> problems, 
>>>>>>>>>>>> not just some.
>>>>>>>>>>>>
>>>>>>>>>>>> On Wednesday, August 3, 2016 at 4:01:15 PM UTC-3, Kevin Liu 
>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>> Hello Community, 
>>>>>>>>>>>>>
>>>>>>>>>>>>> I'm in the last pages of Pedro Domingos' book, the Master 
>>>>>>>>>>>>> Algo, one of two recommended by Bill Gates to learn about AI. 
>>>>>>>>>>>>>
>>>>>>>>>>>>> From the book, I understand all learners have to represent, 
>>>>>>>>>>>>> evaluate, and optimize. There are many types of learners that do 
>>>>>>>>>>>>> this. What 
>>>>>>>>>>>>> Domingos does is generalize these three parts, (1) using Markov 
>>>>>>>>>>>>> Logic 
>>>>>>>>>>>>> Network to represent, (2) posterior probability to evaluate, and 
>>>>>>>>>>>>> (3) 
>>>>>>>>>>>>> genetic search with gradient descent to optimize. The posterior 
>>>>>>>>>>>>> can be 
>>>>>>>>>>>>> replaced for another accuracy measure when it is easier, as 
>>>>>>>>>>>>> genetic search 
>>>>>>>>>>>>> replaced by hill climbing. Where there are 15 popular options for 
>>>>>>>>>>>>> representing, evaluating, and optimizing, Domingos generalized 
>>>>>>>>>>>>> them into 
>>>>>>>>>>>>> three options. The idea is to have one unified learner for any 
>>>>>>>>>>>>> application. 
>>>>>>>>>>>>>
>>>>>>>>>>>>> There is code already done in R 
>>>>>>>>>>>>> https://alchemy.cs.washington.edu/. My question: anybody in 
>>>>>>>>>>>>> the community vested in coding it into Julia?
>>>>>>>>>>>>>
>>>>>>>>>>>>> Thanks. Kevin
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Friday, June 3, 2016 at 3:44:09 PM UTC-3, Kevin Liu wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> https://github.com/tbreloff/OnlineAI.jl/issues/5
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Friday, June 3, 2016 at 11:17:28 AM UTC-3, Kevin Liu wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> I plan to write Julia for the rest of me life... given it 
>>>>>>>>>>>>>>> remains suitable. I am still reading all of Colah's material on 
>>>>>>>>>>>>>>> nets. I ran 
>>>>>>>>>>>>>>> Mocha.jl a couple weeks ago and was very happy to see it work. 
>>>>>>>>>>>>>>> Thanks for 
>>>>>>>>>>>>>>> jumping in and telling me about OnlineAI.jl, I will look into 
>>>>>>>>>>>>>>> it once I am 
>>>>>>>>>>>>>>> ready. From a quick look, perhaps I could help and learn by 
>>>>>>>>>>>>>>> building a very 
>>>>>>>>>>>>>>> clear documentation of it. Would really like to see Julia a 
>>>>>>>>>>>>>>> leap ahead of 
>>>>>>>>>>>>>>> other languages, and plan to contribute heavily to it, but at 
>>>>>>>>>>>>>>> the moment am 
>>>>>>>>>>>>>>> still getting introduced to CS, programming, and nets at the 
>>>>>>>>>>>>>>> basic level. 
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Friday, June 3, 2016 at 10:48:15 AM UTC-3, Tom Breloff 
>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> Kevin: computers that program themselves is a concept which 
>>>>>>>>>>>>>>>> is much closer to reality than most would believe, but 
>>>>>>>>>>>>>>>> julia-users isn't 
>>>>>>>>>>>>>>>> really the best place for this speculation. If you're actually 
>>>>>>>>>>>>>>>> interested 
>>>>>>>>>>>>>>>> in writing code, I'm happy to discuss in OnlineAI.jl. I was 
>>>>>>>>>>>>>>>> thinking about 
>>>>>>>>>>>>>>>> how we might tackle code generation using a neural framework 
>>>>>>>>>>>>>>>> I'm working 
>>>>>>>>>>>>>>>> on. 
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On Friday, June 3, 2016, Kevin Liu <kvt...@gmail.com> 
>>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> If Andrew Ng who cited Gates, and Gates who cited Domingos 
>>>>>>>>>>>>>>>>> (who did not lecture at Google with a TensorFlow question in 
>>>>>>>>>>>>>>>>> the end), were 
>>>>>>>>>>>>>>>>> unsuccessful penny traders, Julia was a language for web 
>>>>>>>>>>>>>>>>> design, and the 
>>>>>>>>>>>>>>>>> tribes in the video didn't actually solve problems, perhaps 
>>>>>>>>>>>>>>>>> this would be a 
>>>>>>>>>>>>>>>>> wildly off-topic, speculative discussion. But these 
>>>>>>>>>>>>>>>>> statements couldn't be 
>>>>>>>>>>>>>>>>> farther from the truth. In fact, if I had known about this 
>>>>>>>>>>>>>>>>> video some 
>>>>>>>>>>>>>>>>> months ago I would've understood better on how to solve a 
>>>>>>>>>>>>>>>>> problem I was 
>>>>>>>>>>>>>>>>> working on.  
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> For the founders of Julia: I understand your tribe is 
>>>>>>>>>>>>>>>>> mainly CS. This master algorithm, as you are aware, would 
>>>>>>>>>>>>>>>>> require 
>>>>>>>>>>>>>>>>> collaboration with other tribes. Just citing the obvious. 
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> On Friday, June 3, 2016 at 10:21:25 AM UTC-3, Kevin Liu 
>>>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> There could be parts missing as Domingos mentions, but 
>>>>>>>>>>>>>>>>>> induction, backpropagation, genetic programming, 
>>>>>>>>>>>>>>>>>> probabilistic inference, 
>>>>>>>>>>>>>>>>>> and SVMs working together-- what's speculative about the 
>>>>>>>>>>>>>>>>>> improved versions 
>>>>>>>>>>>>>>>>>> of these? 
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> Julia was made for AI. Isn't it time for a consolidated 
>>>>>>>>>>>>>>>>>> view on how to reach it? 
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> On Thursday, June 2, 2016 at 11:20:35 PM UTC-3, Isaiah 
>>>>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> This is not a forum for wildly off-topic, speculative 
>>>>>>>>>>>>>>>>>>> discussion.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Take this to Reddit, Hacker News, etc.
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> On Thu, Jun 2, 2016 at 10:01 PM, Kevin Liu <
>>>>>>>>>>>>>>>>>>> kvt...@gmail.com> wrote:
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> I am wondering how Julia fits in with the unified tribes
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> mashable.com/2016/06/01/bill-gates-ai-code-conference/#8VmBFjIiYOqJ
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> https://www.youtube.com/watch?v=B8J4uefCQMc
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>
>>>

Reply via email to