I have no idea where to start and where to finish. Founders' help would be 
wonderful. 

On Tuesday, August 9, 2016 at 12:19:26 AM UTC-3, Kevin Liu wrote:
>
> After which I have to code Felix into Julia, a relational optimizer for 
> statistical inference with Tuffy <http://i.stanford.edu/hazy/tuffy/> 
> inside, for enterprise settings.
>
> On Tuesday, August 9, 2016 at 12:07:32 AM UTC-3, Kevin Liu wrote:
>>
>> Can I get tips on bringing Alchemy's optimized Tuffy 
>> <http://i.stanford.edu/hazy/tuffy/> in Java to Julia while showing the 
>> best of Julia? I am going for the most correct way, even if it means coding 
>> Tuffy into C and Julia.
>>
>> On Sunday, August 7, 2016 at 8:34:37 PM UTC-3, Kevin Liu wrote:
>>>
>>> I'll try to build it, compare it, and show it to you guys. I offered to 
>>> do this as work. I am waiting to see if they will accept it. 
>>>
>>> On Sunday, August 7, 2016 at 6:15:50 PM UTC-3, Stefan Karpinski wrote:
>>>>
>>>> Kevin, as previously requested by Isaiah, please take this to some 
>>>> other forum or maybe start a blog.
>>>>
>>>> On Sat, Aug 6, 2016 at 10:53 PM, Kevin Liu <kvt...@gmail.com> wrote:
>>>>
>>>>> Symmetry-based learning, Domingos, 2014 
>>>>> https://www.microsoft.com/en-us/research/video/symmetry-based-learning/
>>>>>
>>>>> Approach 2: Deep symmetry networks generalize convolutional neural 
>>>>> networks by tying parameters and pooling over an arbitrary symmetry 
>>>>> group, 
>>>>> not just the translation group. In preliminary experiments, they 
>>>>> outperformed convnets on a digit recognition task. 
>>>>>
>>>>> On Friday, August 5, 2016 at 4:56:45 PM UTC-3, Kevin Liu wrote:
>>>>>>
>>>>>> Minsky died of a cerebral hemorrhage at the age of 88.[40] 
>>>>>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-40> Ray 
>>>>>> Kurzweil <https://en.wikipedia.org/wiki/Ray_Kurzweil> says he was 
>>>>>> contacted by the cryonics organization Alcor Life Extension 
>>>>>> Foundation 
>>>>>> <https://en.wikipedia.org/wiki/Alcor_Life_Extension_Foundation> seeking 
>>>>>> Minsky's body.[41] 
>>>>>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-Kurzweil-41> 
>>>>>> Kurzweil 
>>>>>> believes that Minsky was cryonically preserved by Alcor and will be 
>>>>>> revived 
>>>>>> by 2045.[41] 
>>>>>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-Kurzweil-41> 
>>>>>> Minsky 
>>>>>> was a member of Alcor's Scientific Advisory Board 
>>>>>> <https://en.wikipedia.org/wiki/Advisory_Board>.[42] 
>>>>>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-AlcorBoard-42> In 
>>>>>> keeping with their policy of protecting privacy, Alcor will neither 
>>>>>> confirm 
>>>>>> nor deny that Alcor has cryonically preserved Minsky.[43] 
>>>>>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-43> 
>>>>>>
>>>>>> We better do a good job. 
>>>>>>
>>>>>> On Friday, August 5, 2016 at 4:45:42 PM UTC-3, Kevin Liu wrote:
>>>>>>>
>>>>>>> *So, I think in the next 20 years (2003), if we can get rid of all 
>>>>>>> of the traditional approaches to artificial intelligence, like neural 
>>>>>>> nets 
>>>>>>> and genetic algorithms and rule-based systems, and just turn our sights 
>>>>>>> a 
>>>>>>> little bit higher to say, can we make a system that can use all those 
>>>>>>> things for the right kind of problem? Some problems are good for neural 
>>>>>>> nets; we know that others, neural nets are hopeless on them. Genetic 
>>>>>>> algorithms are great for certain things; I suspect I know what they're 
>>>>>>> bad 
>>>>>>> at, and I won't tell you. (Laughter)*  - Minsky, founder of CSAIL 
>>>>>>> MIT
>>>>>>>
>>>>>>> *Those programmers tried to find the single best way to represent 
>>>>>>> knowledge - Only Logic protects us from paradox.* - Minsky (see 
>>>>>>> attachment from his lecture)
>>>>>>>
>>>>>>> On Friday, August 5, 2016 at 8:12:03 AM UTC-3, Kevin Liu wrote:
>>>>>>>>
>>>>>>>> Markov Logic Network is being used for the continuous development 
>>>>>>>> of drugs to cure cancer at MIT's CanceRX <http://cancerx.mit.edu/>, on 
>>>>>>>> DARPA's largest AI project to date, Personalized Assistant that 
>>>>>>>> Learns (PAL) <https://pal.sri.com/>, progenitor of Siri. One of 
>>>>>>>> Alchemy's largest applications to date was to learn a semantic network 
>>>>>>>> (knowledge graph as Google calls it) from the web. 
>>>>>>>>
>>>>>>>> Some on Probabilistic Inductive Logic Programming / Probabilistic 
>>>>>>>> Logic Programming / Statistical Relational Learning from CSAIL 
>>>>>>>> <http://people.csail.mit.edu/kersting/ecmlpkdd05_pilp/pilp_ida2005_tut.pdf>
>>>>>>>>  (my 
>>>>>>>> understanding is Alchemy does PILP from entailment, proofs, and 
>>>>>>>> interpretation)
>>>>>>>>
>>>>>>>> The MIT Probabilistic Computing Project (where there is Picture, an 
>>>>>>>> extension of Julia, for computer vision; Watch the video from Vikash) 
>>>>>>>> <http://probcomp.csail.mit.edu/index.html>
>>>>>>>>
>>>>>>>> Probabilistic programming could do for Bayesian ML what Theano has 
>>>>>>>> done for neural networks. 
>>>>>>>> <http://www.inference.vc/deep-learning-is-easy/> - Ferenc Huszár
>>>>>>>>
>>>>>>>> Picture doesn't appear to be open-source, even though its Paper is 
>>>>>>>> available. 
>>>>>>>>
>>>>>>>> I'm in the process of comparing the Picture Paper and Alchemy code 
>>>>>>>> and would like to have an open-source PILP from Julia that combines 
>>>>>>>> the 
>>>>>>>> best of both. 
>>>>>>>>
>>>>>>>> On Wednesday, August 3, 2016 at 5:01:02 PM UTC-3, Christof Stocker 
>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>> This sounds like it could be a great contribution. I shall keep a 
>>>>>>>>> curious eye on your progress
>>>>>>>>>
>>>>>>>>> Am Mittwoch, 3. August 2016 21:53:54 UTC+2 schrieb Kevin Liu:
>>>>>>>>>>
>>>>>>>>>> Thanks for the advice Cristof. I am only interested in people 
>>>>>>>>>> wanting to code it in Julia, from R by Domingos. The algo has been 
>>>>>>>>>> successfully applied in many areas, even though there are many other 
>>>>>>>>>> areas 
>>>>>>>>>> remaining. 
>>>>>>>>>>
>>>>>>>>>> On Wed, Aug 3, 2016 at 4:45 PM, Christof Stocker <
>>>>>>>>>> stocker....@gmail.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> Hello Kevin,
>>>>>>>>>>>
>>>>>>>>>>> Enthusiasm is a good thing and you should hold on to that. But 
>>>>>>>>>>> to save yourself some headache or disappointment down the road I 
>>>>>>>>>>> advice a 
>>>>>>>>>>> level head. Nothing is really as bluntly obviously solved as it may 
>>>>>>>>>>> seems 
>>>>>>>>>>> at first glance after listening to brilliant people explain things. 
>>>>>>>>>>> A 
>>>>>>>>>>> physics professor of mine once told me that one of the (he thinks) 
>>>>>>>>>>> most 
>>>>>>>>>>> malicious factors to his past students progress where overstated 
>>>>>>>>>>> results/conclusions by other researches (such as premature 
>>>>>>>>>>> announcements 
>>>>>>>>>>> from CERN). I am no mathematician, but as far as I can judge is the 
>>>>>>>>>>> no free 
>>>>>>>>>>> lunch theorem of pure mathematical nature and not something induced 
>>>>>>>>>>> empirically. These kind of results are not that easily to get rid 
>>>>>>>>>>> of. If 
>>>>>>>>>>> someone (especially an expert) states such a theorem will prove 
>>>>>>>>>>> wrong I 
>>>>>>>>>>> would be inclined to believe that he is not talking about 
>>>>>>>>>>> literally, but 
>>>>>>>>>>> instead is just trying to make a point about a more or less 
>>>>>>>>>>> practical 
>>>>>>>>>>> implication.
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> Am Mittwoch, 3. August 2016 21:27:05 UTC+2 schrieb Kevin Liu:
>>>>>>>>>>>>
>>>>>>>>>>>> The Markov logic network represents a probability distribution 
>>>>>>>>>>>> over the states of a complex system (i.e. a cell), comprised of 
>>>>>>>>>>>> entities, 
>>>>>>>>>>>> where logic formulas encode the dependencies between them. 
>>>>>>>>>>>>
>>>>>>>>>>>> On Wednesday, August 3, 2016 at 4:19:09 PM UTC-3, Kevin Liu 
>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>> Alchemy is like an inductive Turing machine, to be programmed 
>>>>>>>>>>>>> to learn broadly or restrictedly.
>>>>>>>>>>>>>
>>>>>>>>>>>>> The logic formulas from rules through which it represents can 
>>>>>>>>>>>>> be inconsistent, incomplete, or even incorrect-- the learning and 
>>>>>>>>>>>>> probabilistic reasoning will correct them. The key point is that 
>>>>>>>>>>>>> Alchemy 
>>>>>>>>>>>>> doesn't have to learn from scratch, proving Wolpert and 
>>>>>>>>>>>>> Macready's no free 
>>>>>>>>>>>>> lunch theorem wrong by performing well on a variety of classes of 
>>>>>>>>>>>>> problems, 
>>>>>>>>>>>>> not just some.
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Wednesday, August 3, 2016 at 4:01:15 PM UTC-3, Kevin Liu 
>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Hello Community, 
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> I'm in the last pages of Pedro Domingos' book, the Master 
>>>>>>>>>>>>>> Algo, one of two recommended by Bill Gates to learn about AI. 
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> From the book, I understand all learners have to represent, 
>>>>>>>>>>>>>> evaluate, and optimize. There are many types of learners that do 
>>>>>>>>>>>>>> this. What 
>>>>>>>>>>>>>> Domingos does is generalize these three parts, (1) using Markov 
>>>>>>>>>>>>>> Logic 
>>>>>>>>>>>>>> Network to represent, (2) posterior probability to evaluate, and 
>>>>>>>>>>>>>> (3) 
>>>>>>>>>>>>>> genetic search with gradient descent to optimize. The posterior 
>>>>>>>>>>>>>> can be 
>>>>>>>>>>>>>> replaced for another accuracy measure when it is easier, as 
>>>>>>>>>>>>>> genetic search 
>>>>>>>>>>>>>> replaced by hill climbing. Where there are 15 popular options 
>>>>>>>>>>>>>> for 
>>>>>>>>>>>>>> representing, evaluating, and optimizing, Domingos generalized 
>>>>>>>>>>>>>> them into 
>>>>>>>>>>>>>> three options. The idea is to have one unified learner for any 
>>>>>>>>>>>>>> application. 
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> There is code already done in R 
>>>>>>>>>>>>>> https://alchemy.cs.washington.edu/. My question: anybody in 
>>>>>>>>>>>>>> the community vested in coding it into Julia?
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Thanks. Kevin
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Friday, June 3, 2016 at 3:44:09 PM UTC-3, Kevin Liu wrote:
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> https://github.com/tbreloff/OnlineAI.jl/issues/5
>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>> On Friday, June 3, 2016 at 11:17:28 AM UTC-3, Kevin Liu 
>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> I plan to write Julia for the rest of me life... given it 
>>>>>>>>>>>>>>>> remains suitable. I am still reading all of Colah's material 
>>>>>>>>>>>>>>>> on nets. I ran 
>>>>>>>>>>>>>>>> Mocha.jl a couple weeks ago and was very happy to see it work. 
>>>>>>>>>>>>>>>> Thanks for 
>>>>>>>>>>>>>>>> jumping in and telling me about OnlineAI.jl, I will look into 
>>>>>>>>>>>>>>>> it once I am 
>>>>>>>>>>>>>>>> ready. From a quick look, perhaps I could help and learn by 
>>>>>>>>>>>>>>>> building a very 
>>>>>>>>>>>>>>>> clear documentation of it. Would really like to see Julia a 
>>>>>>>>>>>>>>>> leap ahead of 
>>>>>>>>>>>>>>>> other languages, and plan to contribute heavily to it, but at 
>>>>>>>>>>>>>>>> the moment am 
>>>>>>>>>>>>>>>> still getting introduced to CS, programming, and nets at the 
>>>>>>>>>>>>>>>> basic level. 
>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>> On Friday, June 3, 2016 at 10:48:15 AM UTC-3, Tom Breloff 
>>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> Kevin: computers that program themselves is a concept 
>>>>>>>>>>>>>>>>> which is much closer to reality than most would believe, but 
>>>>>>>>>>>>>>>>> julia-users 
>>>>>>>>>>>>>>>>> isn't really the best place for this speculation. If you're 
>>>>>>>>>>>>>>>>> actually 
>>>>>>>>>>>>>>>>> interested in writing code, I'm happy to discuss in 
>>>>>>>>>>>>>>>>> OnlineAI.jl. I was 
>>>>>>>>>>>>>>>>> thinking about how we might tackle code generation using a 
>>>>>>>>>>>>>>>>> neural framework 
>>>>>>>>>>>>>>>>> I'm working on. 
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>> On Friday, June 3, 2016, Kevin Liu <kvt...@gmail.com> 
>>>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> If Andrew Ng who cited Gates, and Gates who cited 
>>>>>>>>>>>>>>>>>> Domingos (who did not lecture at Google with a TensorFlow 
>>>>>>>>>>>>>>>>>> question in the 
>>>>>>>>>>>>>>>>>> end), were unsuccessful penny traders, Julia was a language 
>>>>>>>>>>>>>>>>>> for web design, 
>>>>>>>>>>>>>>>>>> and the tribes in the video didn't actually solve problems, 
>>>>>>>>>>>>>>>>>> perhaps this 
>>>>>>>>>>>>>>>>>> would be a wildly off-topic, speculative discussion. But 
>>>>>>>>>>>>>>>>>> these statements 
>>>>>>>>>>>>>>>>>> couldn't be farther from the truth. In fact, if I had known 
>>>>>>>>>>>>>>>>>> about this 
>>>>>>>>>>>>>>>>>> video some months ago I would've understood better on how to 
>>>>>>>>>>>>>>>>>> solve a 
>>>>>>>>>>>>>>>>>> problem I was working on.  
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> For the founders of Julia: I understand your tribe is 
>>>>>>>>>>>>>>>>>> mainly CS. This master algorithm, as you are aware, would 
>>>>>>>>>>>>>>>>>> require 
>>>>>>>>>>>>>>>>>> collaboration with other tribes. Just citing the obvious. 
>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>> On Friday, June 3, 2016 at 10:21:25 AM UTC-3, Kevin Liu 
>>>>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> There could be parts missing as Domingos mentions, but 
>>>>>>>>>>>>>>>>>>> induction, backpropagation, genetic programming, 
>>>>>>>>>>>>>>>>>>> probabilistic inference, 
>>>>>>>>>>>>>>>>>>> and SVMs working together-- what's speculative about the 
>>>>>>>>>>>>>>>>>>> improved versions 
>>>>>>>>>>>>>>>>>>> of these? 
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> Julia was made for AI. Isn't it time for a consolidated 
>>>>>>>>>>>>>>>>>>> view on how to reach it? 
>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>> On Thursday, June 2, 2016 at 11:20:35 PM UTC-3, Isaiah 
>>>>>>>>>>>>>>>>>>> wrote:
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> This is not a forum for wildly off-topic, speculative 
>>>>>>>>>>>>>>>>>>>> discussion.
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> Take this to Reddit, Hacker News, etc.
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>> On Thu, Jun 2, 2016 at 10:01 PM, Kevin Liu <
>>>>>>>>>>>>>>>>>>>> kvt...@gmail.com> wrote:
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> I am wondering how Julia fits in with the unified 
>>>>>>>>>>>>>>>>>>>>> tribes
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> mashable.com/2016/06/01/bill-gates-ai-code-conference/#8VmBFjIiYOqJ
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>> https://www.youtube.com/watch?v=B8J4uefCQMc
>>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>>>>>>>>>>>
>>>>>>>>>>
>>>>

Reply via email to