Two things. One, yes I am using a form of relational learning on that project. Interesting stuff is coming out of it. But two, I was just trying to help. You're clearly abusing the mailing list and are probably going to get banned if you don't stop. I suggest blogging these musings, and keeping the mailing list for discussions about Julia code.
On Friday, September 2, 2016 at 7:23:27 AM UTC-7, Kevin Liu wrote: > > May I also point out to the My settings button on your top right corner > > My topic email subscriptions > Unsubscribe from this thread, which would've > spared you the message. > > On Friday, September 2, 2016 at 11:19:42 AM UTC-3, Kevin Liu wrote: >> >> Hello Chris. Have you been applying relational learning to your Neural >> Crest Migration Patterns in Craniofacial Development research project? It >> could enhance your insights. >> >> On Friday, September 2, 2016 at 6:18:15 AM UTC-3, Chris Rackauckas wrote: >>> >>> This entire thread is a trip... a trip which is not really relevant to >>> julia-users. You may want to share these musings in the form of a blog >>> instead of posting them here. >>> >>> On Friday, September 2, 2016 at 1:41:03 AM UTC-7, Kevin Liu wrote: >>>> >>>> Princeton's post: >>>> http://www.nytimes.com/2016/08/28/world/europe/france-burkini-bikini-ban.html?_r=1 >>>> >>>> Only logic saves us from paradox. - Minsky >>>> >>>> On Thursday, August 25, 2016 at 10:18:27 PM UTC-3, Kevin Liu wrote: >>>>> >>>>> Tim Holy, I am watching your keynote speech at JuliaCon 2016 where you >>>>> mention the best optimization is not doing the computation at all. >>>>> >>>>> Domingos talks about that in his book, where an efficient kind of >>>>> learning is by analogy, with no model at all, and how numerous scientific >>>>> discoveries have been made that way, e.g. Bohr's analogy of the solar >>>>> system to the atom. Analogizers learn by hypothesizing that entities with >>>>> similar known properties have similar unknown ones. >>>>> >>>>> MLN can reproduce structure mapping, which is the more powerful type >>>>> of analogy, that can make inferences from one domain (solar system) to >>>>> another (atom). This can be done by learning formulas that don't refer to >>>>> any of the specific relations in the source domain (general formulas). >>>>> >>>>> Seth and Tim have been helping me a lot with putting the pieces >>>>> together for MLN in the repo I created >>>>> <https://github.com/hpoit/Kenya.jl/issues/2>, and more help is always >>>>> welcome. I would like to write MLN in idiomatic Julia. My question at the >>>>> moment to you and the community is how to keep mappings of first-order >>>>> harmonic functions type-stable in Julia? I am just getting acquainted >>>>> with >>>>> the type field. >>>>> >>>>> On Tuesday, August 9, 2016 at 9:02:25 AM UTC-3, Kevin Liu wrote: >>>>>> >>>>>> Helping me separate the process in parts and priorities would be a >>>>>> lot of help. >>>>>> >>>>>> On Tuesday, August 9, 2016 at 8:41:03 AM UTC-3, Kevin Liu wrote: >>>>>>> >>>>>>> Tim Holy, what if I could tap into the well of knowledge that you >>>>>>> are to speed up things? Can you imagine if every learner had to start >>>>>>> without priors? >>>>>>> >>>>>>> > On Aug 9, 2016, at 07:06, Tim Holy <[email protected]> wrote: >>>>>>> > >>>>>>> > I'd recommend starting by picking a very small project. For >>>>>>> example, fix a bug >>>>>>> > or implement a small improvement in a package that you already >>>>>>> find useful or >>>>>>> > interesting. That way you'll get some guidance while making a >>>>>>> positive >>>>>>> > contribution; once you know more about julia, it will be easier to >>>>>>> see your >>>>>>> > way forward. >>>>>>> > >>>>>>> > Best, >>>>>>> > --Tim >>>>>>> > >>>>>>> >> On Monday, August 8, 2016 8:22:01 PM CDT Kevin Liu wrote: >>>>>>> >> I have no idea where to start and where to finish. Founders' help >>>>>>> would be >>>>>>> >> wonderful. >>>>>>> >> >>>>>>> >>> On Tuesday, August 9, 2016 at 12:19:26 AM UTC-3, Kevin Liu >>>>>>> wrote: >>>>>>> >>> After which I have to code Felix into Julia, a relational >>>>>>> optimizer for >>>>>>> >>> statistical inference with Tuffy < >>>>>>> http://i.stanford.edu/hazy/tuffy/> >>>>>>> >>> inside, for enterprise settings. >>>>>>> >>> >>>>>>> >>>> On Tuesday, August 9, 2016 at 12:07:32 AM UTC-3, Kevin Liu >>>>>>> wrote: >>>>>>> >>>> Can I get tips on bringing Alchemy's optimized Tuffy >>>>>>> >>>> <http://i.stanford.edu/hazy/tuffy/> in Java to Julia while >>>>>>> showing the >>>>>>> >>>> best of Julia? I am going for the most correct way, even if it >>>>>>> means >>>>>>> >>>> coding >>>>>>> >>>> Tuffy into C and Julia. >>>>>>> >>>> >>>>>>> >>>>> On Sunday, August 7, 2016 at 8:34:37 PM UTC-3, Kevin Liu >>>>>>> wrote: >>>>>>> >>>>> I'll try to build it, compare it, and show it to you guys. I >>>>>>> offered to >>>>>>> >>>>> do this as work. I am waiting to see if they will accept it. >>>>>>> >>>>> >>>>>>> >>>>>> On Sunday, August 7, 2016 at 6:15:50 PM UTC-3, Stefan >>>>>>> Karpinski wrote: >>>>>>> >>>>>> Kevin, as previously requested by Isaiah, please take this to >>>>>>> some >>>>>>> >>>>>> other forum or maybe start a blog. >>>>>>> >>>>>> >>>>>>> >>>>>>> On Sat, Aug 6, 2016 at 10:53 PM, Kevin Liu <[email protected]> >>>>>>> wrote: >>>>>>> >>>>>>> Symmetry-based learning, Domingos, 2014 >>>>>>> >>>>>>> >>>>>>> https://www.microsoft.com/en-us/research/video/symmetry-based-learning >>>>>>> >>>>>>> / >>>>>>> >>>>>>> >>>>>>> >>>>>>> Approach 2: Deep symmetry networks generalize convolutional >>>>>>> neural >>>>>>> >>>>>>> networks by tying parameters and pooling over an arbitrary >>>>>>> symmetry >>>>>>> >>>>>>> group, >>>>>>> >>>>>>> not just the translation group. In preliminary experiments, >>>>>>> they >>>>>>> >>>>>>> outperformed convnets on a digit recognition task. >>>>>>> >>>>>>> >>>>>>> >>>>>>>> On Friday, August 5, 2016 at 4:56:45 PM UTC-3, Kevin Liu >>>>>>> wrote: >>>>>>> >>>>>>>> Minsky died of a cerebral hemorrhage at the age of 88.[40] >>>>>>> >>>>>>>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-40> >>>>>>> Ray >>>>>>> >>>>>>>> Kurzweil <https://en.wikipedia.org/wiki/Ray_Kurzweil> says >>>>>>> he was >>>>>>> >>>>>>>> contacted by the cryonics organization Alcor Life Extension >>>>>>> >>>>>>>> Foundation >>>>>>> >>>>>>>> < >>>>>>> https://en.wikipedia.org/wiki/Alcor_Life_Extension_Foundation> >>>>>>> >>>>>>>> seeking >>>>>>> >>>>>>>> Minsky's body.[41] >>>>>>> >>>>>>>> < >>>>>>> https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-Kurzweil-41> >>>>>>> >>>>>>>> Kurzweil believes that Minsky was cryonically preserved by >>>>>>> Alcor and >>>>>>> >>>>>>>> will be revived by 2045.[41] >>>>>>> >>>>>>>> < >>>>>>> https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-Kurzweil-41> >>>>>>> >>>>>>>> Minsky >>>>>>> >>>>>>>> was a member of Alcor's Scientific Advisory Board >>>>>>> >>>>>>>> <https://en.wikipedia.org/wiki/Advisory_Board>.[42] >>>>>>> >>>>>>>> < >>>>>>> https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-AlcorBoard-42> >>>>>>> >>>>>>> >>>>>>>> In >>>>>>> >>>>>>>> keeping with their policy of protecting privacy, Alcor will >>>>>>> neither >>>>>>> >>>>>>>> confirm >>>>>>> >>>>>>>> nor deny that Alcor has cryonically preserved Minsky.[43] >>>>>>> >>>>>>>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-43> >>>>>>> >>>>>>>> >>>>>>> >>>>>>>> We better do a good job. >>>>>>> >>>>>>>> >>>>>>> >>>>>>>>> On Friday, August 5, 2016 at 4:45:42 PM UTC-3, Kevin Liu >>>>>>> wrote: >>>>>>> >>>>>>>>> *So, I think in the next 20 years (2003), if we can get >>>>>>> rid of all >>>>>>> >>>>>>>>> of the traditional approaches to artificial intelligence, >>>>>>> like >>>>>>> >>>>>>>>> neural nets >>>>>>> >>>>>>>>> and genetic algorithms and rule-based systems, and just >>>>>>> turn our >>>>>>> >>>>>>>>> sights a >>>>>>> >>>>>>>>> little bit higher to say, can we make a system that can >>>>>>> use all >>>>>>> >>>>>>>>> those >>>>>>> >>>>>>>>> things for the right kind of problem? Some problems are >>>>>>> good for >>>>>>> >>>>>>>>> neural >>>>>>> >>>>>>>>> nets; we know that others, neural nets are hopeless on >>>>>>> them. Genetic >>>>>>> >>>>>>>>> algorithms are great for certain things; I suspect I know >>>>>>> what >>>>>>> >>>>>>>>> they're bad >>>>>>> >>>>>>>>> at, and I won't tell you. (Laughter)* - Minsky, founder >>>>>>> of CSAIL >>>>>>> >>>>>>>>> MIT >>>>>>> >>>>>>>>> >>>>>>> >>>>>>>>> *Those programmers tried to find the single best way to >>>>>>> represent >>>>>>> >>>>>>>>> knowledge - Only Logic protects us from paradox.* - Minsky >>>>>>> (see >>>>>>> >>>>>>>>> attachment from his lecture) >>>>>>> >>>>>>>>> >>>>>>> >>>>>>>>>> On Friday, August 5, 2016 at 8:12:03 AM UTC-3, Kevin Liu >>>>>>> wrote: >>>>>>> >>>>>>>>>> Markov Logic Network is being used for the continuous >>>>>>> development >>>>>>> >>>>>>>>>> of drugs to cure cancer at MIT's CanceRX < >>>>>>> http://cancerx.mit.edu/>, >>>>>>> >>>>>>>>>> on >>>>>>> >>>>>>>>>> DARPA's largest AI project to date, Personalized >>>>>>> Assistant that >>>>>>> >>>>>>>>>> Learns (PAL) <https://pal.sri.com/>, progenitor of Siri. >>>>>>> One of >>>>>>> >>>>>>>>>> Alchemy's largest applications to date was to learn a >>>>>>> semantic >>>>>>> >>>>>>>>>> network >>>>>>> >>>>>>>>>> (knowledge graph as Google calls it) from the web. >>>>>>> >>>>>>>>>> >>>>>>> >>>>>>>>>> Some on Probabilistic Inductive Logic Programming / >>>>>>> Probabilistic >>>>>>> >>>>>>>>>> Logic Programming / Statistical Relational Learning from >>>>>>> CSAIL >>>>>>> >>>>>>>>>> < >>>>>>> http://people.csail.mit.edu/kersting/ecmlpkdd05_pilp/pilp_ida2005_ >>>>>>> >>>>>>>>>> tut.pdf> (my understanding is Alchemy does PILP from >>>>>>> entailment, >>>>>>> >>>>>>>>>> proofs, and >>>>>>> >>>>>>>>>> interpretation) >>>>>>> >>>>>>>>>> >>>>>>> >>>>>>>>>> The MIT Probabilistic Computing Project (where there is >>>>>>> Picture, an >>>>>>> >>>>>>>>>> extension of Julia, for computer vision; Watch the video >>>>>>> from >>>>>>> >>>>>>>>>> Vikash) >>>>>>> >>>>>>>>>> <http://probcomp.csail.mit.edu/index.html> >>>>>>> >>>>>>>>>> >>>>>>> >>>>>>>>>> Probabilistic programming could do for Bayesian ML what >>>>>>> Theano has >>>>>>> >>>>>>>>>> done for neural networks. >>>>>>> >>>>>>>>>> <http://www.inference.vc/deep-learning-is-easy/> - >>>>>>> Ferenc Huszár >>>>>>> >>>>>>>>>> >>>>>>> >>>>>>>>>> Picture doesn't appear to be open-source, even though its >>>>>>> Paper is >>>>>>> >>>>>>>>>> available. >>>>>>> >>>>>>>>>> >>>>>>> >>>>>>>>>> I'm in the process of comparing the Picture Paper and >>>>>>> Alchemy code >>>>>>> >>>>>>>>>> and would like to have an open-source PILP from Julia >>>>>>> that combines >>>>>>> >>>>>>>>>> the >>>>>>> >>>>>>>>>> best of both. >>>>>>> >>>>>>>>>> >>>>>>> >>>>>>>>>> On Wednesday, August 3, 2016 at 5:01:02 PM UTC-3, >>>>>>> Christof Stocker >>>>>>> >>>>>>>>>> >>>>>>> >>>>>>>>>> wrote: >>>>>>> >>>>>>>>>>> This sounds like it could be a great contribution. I >>>>>>> shall keep a >>>>>>> >>>>>>>>>>> curious eye on your progress >>>>>>> >>>>>>>>>>> >>>>>>> >>>>>>>>>>> Am Mittwoch, 3. August 2016 21:53:54 UTC+2 schrieb Kevin >>>>>>> Liu: >>>>>>> >>>>>>>>>>>> Thanks for the advice Cristof. I am only interested in >>>>>>> people >>>>>>> >>>>>>>>>>>> wanting to code it in Julia, from R by Domingos. The >>>>>>> algo has >>>>>>> >>>>>>>>>>>> been >>>>>>> >>>>>>>>>>>> successfully applied in many areas, even though there >>>>>>> are many >>>>>>> >>>>>>>>>>>> other areas >>>>>>> >>>>>>>>>>>> remaining. >>>>>>> >>>>>>>>>>>> >>>>>>> >>>>>>>>>>>> On Wed, Aug 3, 2016 at 4:45 PM, Christof Stocker < >>>>>>> >>>>>>>>>>>> >>>>>>> >>>>>>>>>>>> [email protected]> wrote: >>>>>>> >>>>>>>>>>>>> Hello Kevin, >>>>>>> >>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>>> Enthusiasm is a good thing and you should hold on to >>>>>>> that. But >>>>>>> >>>>>>>>>>>>> to save yourself some headache or disappointment down >>>>>>> the road I >>>>>>> >>>>>>>>>>>>> advice a >>>>>>> >>>>>>>>>>>>> level head. Nothing is really as bluntly obviously >>>>>>> solved as it >>>>>>> >>>>>>>>>>>>> may seems >>>>>>> >>>>>>>>>>>>> at first glance after listening to brilliant people >>>>>>> explain >>>>>>> >>>>>>>>>>>>> things. A >>>>>>> >>>>>>>>>>>>> physics professor of mine once told me that one of the >>>>>>> (he >>>>>>> >>>>>>>>>>>>> thinks) most >>>>>>> >>>>>>>>>>>>> malicious factors to his past students progress where >>>>>>> overstated >>>>>>> >>>>>>>>>>>>> results/conclusions by other researches (such as >>>>>>> premature >>>>>>> >>>>>>>>>>>>> announcements >>>>>>> >>>>>>>>>>>>> from CERN). I am no mathematician, but as far as I can >>>>>>> judge is >>>>>>> >>>>>>>>>>>>> the no free >>>>>>> >>>>>>>>>>>>> lunch theorem of pure mathematical nature and not >>>>>>> something >>>>>>> >>>>>>>>>>>>> induced >>>>>>> >>>>>>>>>>>>> empirically. These kind of results are not that easily >>>>>>> to get >>>>>>> >>>>>>>>>>>>> rid of. If >>>>>>> >>>>>>>>>>>>> someone (especially an expert) states such a theorem >>>>>>> will prove >>>>>>> >>>>>>>>>>>>> wrong I >>>>>>> >>>>>>>>>>>>> would be inclined to believe that he is not talking >>>>>>> about >>>>>>> >>>>>>>>>>>>> literally, but >>>>>>> >>>>>>>>>>>>> instead is just trying to make a point about a more or >>>>>>> less >>>>>>> >>>>>>>>>>>>> practical >>>>>>> >>>>>>>>>>>>> implication. >>>>>>> >>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>>> Am Mittwoch, 3. August 2016 21:27:05 UTC+2 schrieb >>>>>>> Kevin Liu: >>>>>>> >>>>>>>>>>>>>> The Markov logic network represents a probability >>>>>>> distribution >>>>>>> >>>>>>>>>>>>>> over the states of a complex system (i.e. a cell), >>>>>>> comprised of >>>>>>> >>>>>>>>>>>>>> entities, >>>>>>> >>>>>>>>>>>>>> where logic formulas encode the dependencies between >>>>>>> them. >>>>>>> >>>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>>>> On Wednesday, August 3, 2016 at 4:19:09 PM UTC-3, >>>>>>> Kevin Liu >>>>>>> >>>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>>>> wrote: >>>>>>> >>>>>>>>>>>>>>> Alchemy is like an inductive Turing machine, to be >>>>>>> programmed >>>>>>> >>>>>>>>>>>>>>> to learn broadly or restrictedly. >>>>>>> >>>>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>>>>> The logic formulas from rules through which it >>>>>>> represents can >>>>>>> >>>>>>>>>>>>>>> be inconsistent, incomplete, or even incorrect-- the >>>>>>> learning >>>>>>> >>>>>>>>>>>>>>> and >>>>>>> >>>>>>>>>>>>>>> probabilistic reasoning will correct them. The key >>>>>>> point is >>>>>>> >>>>>>>>>>>>>>> that Alchemy >>>>>>> >>>>>>>>>>>>>>> doesn't have to learn from scratch, proving Wolpert >>>>>>> and >>>>>>> >>>>>>>>>>>>>>> Macready's no free >>>>>>> >>>>>>>>>>>>>>> lunch theorem wrong by performing well on a variety >>>>>>> of classes >>>>>>> >>>>>>>>>>>>>>> of problems, >>>>>>> >>>>>>>>>>>>>>> not just some. >>>>>>> >>>>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>>>>> On Wednesday, August 3, 2016 at 4:01:15 PM UTC-3, >>>>>>> Kevin Liu >>>>>>> >>>>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>>>>> wrote: >>>>>>> >>>>>>>>>>>>>>>> Hello Community, >>>>>>> >>>>>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>>>>>> I'm in the last pages of Pedro Domingos' book, the >>>>>>> Master >>>>>>> >>>>>>>>>>>>>>>> Algo, one of two recommended by Bill Gates to learn >>>>>>> about AI. >>>>>>> >>>>>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>>>>>> From the book, I understand all learners have to >>>>>>> represent, >>>>>>> >>>>>>>>>>>>>>>> evaluate, and optimize. There are many types of >>>>>>> learners that >>>>>>> >>>>>>>>>>>>>>>> do this. What >>>>>>> >>>>>>>>>>>>>>>> Domingos does is generalize these three parts, (1) >>>>>>> using >>>>>>> >>>>>>>>>>>>>>>> Markov Logic >>>>>>> >>>>>>>>>>>>>>>> Network to represent, (2) posterior probability to >>>>>>> evaluate, >>>>>>> >>>>>>>>>>>>>>>> and (3) >>>>>>> >>>>>>>>>>>>>>>> genetic search with gradient descent to optimize. >>>>>>> The >>>>>>> >>>>>>>>>>>>>>>> posterior can be >>>>>>> >>>>>>>>>>>>>>>> replaced for another accuracy measure when it is >>>>>>> easier, as >>>>>>> >>>>>>>>>>>>>>>> genetic search >>>>>>> >>>>>>>>>>>>>>>> replaced by hill climbing. Where there are 15 >>>>>>> popular options >>>>>>> >>>>>>>>>>>>>>>> for >>>>>>> >>>>>>>>>>>>>>>> representing, evaluating, and optimizing, Domingos >>>>>>> >>>>>>>>>>>>>>>> generalized them into >>>>>>> >>>>>>>>>>>>>>>> three options. The idea is to have one unified >>>>>>> learner for >>>>>>> >>>>>>>>>>>>>>>> any application. >>>>>>> >>>>>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>>>>>> There is code already done in R >>>>>>> >>>>>>>>>>>>>>>> https://alchemy.cs.washington.edu/. My question: >>>>>>> anybody in >>>>>>> >>>>>>>>>>>>>>>> the community vested in coding it into Julia? >>>>>>> >>>>>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>>>>>> Thanks. Kevin >>>>>>> >>>>>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>>>>>>> On Friday, June 3, 2016 at 3:44:09 PM UTC-3, Kevin >>>>>>> Liu wrote: >>>>>>> >>>>>>>>>>>>>>>>> https://github.com/tbreloff/OnlineAI.jl/issues/5 >>>>>>> >>>>>>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>>>>>>> On Friday, June 3, 2016 at 11:17:28 AM UTC-3, >>>>>>> Kevin Liu >>>>>>> >>>>>>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>>>>>>> wrote: >>>>>>> >>>>>>>>>>>>>>>>>> I plan to write Julia for the rest of me life... >>>>>>> given it >>>>>>> >>>>>>>>>>>>>>>>>> remains suitable. I am still reading all of >>>>>>> Colah's >>>>>>> >>>>>>>>>>>>>>>>>> material on nets. I ran >>>>>>> >>>>>>>>>>>>>>>>>> Mocha.jl a couple weeks ago and was very happy to >>>>>>> see it >>>>>>> >>>>>>>>>>>>>>>>>> work. Thanks for >>>>>>> >>>>>>>>>>>>>>>>>> jumping in and telling me about OnlineAI.jl, I >>>>>>> will look >>>>>>> >>>>>>>>>>>>>>>>>> into it once I am >>>>>>> >>>>>>>>>>>>>>>>>> ready. From a quick look, perhaps I could help >>>>>>> and learn by >>>>>>> >>>>>>>>>>>>>>>>>> building a very >>>>>>> >>>>>>>>>>>>>>>>>> clear documentation of it. Would really like to >>>>>>> see Julia a >>>>>>> >>>>>>>>>>>>>>>>>> leap ahead of >>>>>>> >>>>>>>>>>>>>>>>>> other languages, and plan to contribute heavily >>>>>>> to it, but >>>>>>> >>>>>>>>>>>>>>>>>> at the moment am >>>>>>> >>>>>>>>>>>>>>>>>> still getting introduced to CS, programming, and >>>>>>> nets at >>>>>>> >>>>>>>>>>>>>>>>>> the basic level. >>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>>>>>>>> On Friday, June 3, 2016 at 10:48:15 AM UTC-3, Tom >>>>>>> Breloff >>>>>>> >>>>>>>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>>>>>>>> wrote: >>>>>>> >>>>>>>>>>>>>>>>>>> Kevin: computers that program themselves is a >>>>>>> concept >>>>>>> >>>>>>>>>>>>>>>>>>> which is much closer to reality than most would >>>>>>> believe, >>>>>>> >>>>>>>>>>>>>>>>>>> but julia-users >>>>>>> >>>>>>>>>>>>>>>>>>> isn't really the best place for this >>>>>>> speculation. If >>>>>>> >>>>>>>>>>>>>>>>>>> you're actually >>>>>>> >>>>>>>>>>>>>>>>>>> interested in writing code, I'm happy to discuss >>>>>>> in >>>>>>> >>>>>>>>>>>>>>>>>>> OnlineAI.jl. I was >>>>>>> >>>>>>>>>>>>>>>>>>> thinking about how we might tackle code >>>>>>> generation using a >>>>>>> >>>>>>>>>>>>>>>>>>> neural framework >>>>>>> >>>>>>>>>>>>>>>>>>> I'm working on. >>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>>>>>>>>> On Friday, June 3, 2016, Kevin Liu < >>>>>>> [email protected]> >>>>>>> >>>>>>>>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>>>>>>>>> wrote: >>>>>>> >>>>>>>>>>>>>>>>>>>> If Andrew Ng who cited Gates, and Gates who >>>>>>> cited >>>>>>> >>>>>>>>>>>>>>>>>>>> Domingos (who did not lecture at Google with a >>>>>>> TensorFlow >>>>>>> >>>>>>>>>>>>>>>>>>>> question in the >>>>>>> >>>>>>>>>>>>>>>>>>>> end), were unsuccessful penny traders, Julia >>>>>>> was a >>>>>>> >>>>>>>>>>>>>>>>>>>> language for web design, >>>>>>> >>>>>>>>>>>>>>>>>>>> and the tribes in the video didn't actually >>>>>>> solve >>>>>>> >>>>>>>>>>>>>>>>>>>> problems, perhaps this >>>>>>> >>>>>>>>>>>>>>>>>>>> would be a wildly off-topic, speculative >>>>>>> discussion. But >>>>>>> >>>>>>>>>>>>>>>>>>>> these statements >>>>>>> >>>>>>>>>>>>>>>>>>>> couldn't be farther from the truth. In fact, if >>>>>>> I had >>>>>>> >>>>>>>>>>>>>>>>>>>> known about this >>>>>>> >>>>>>>>>>>>>>>>>>>> video some months ago I would've understood >>>>>>> better on how >>>>>>> >>>>>>>>>>>>>>>>>>>> to solve a >>>>>>> >>>>>>>>>>>>>>>>>>>> problem I was working on. >>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>>>>>>>>>> For the founders of Julia: I understand your >>>>>>> tribe is >>>>>>> >>>>>>>>>>>>>>>>>>>> mainly CS. This master algorithm, as you are >>>>>>> aware, would >>>>>>> >>>>>>>>>>>>>>>>>>>> require >>>>>>> >>>>>>>>>>>>>>>>>>>> collaboration with other tribes. Just citing >>>>>>> the obvious. >>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>>>>>>>>>> On Friday, June 3, 2016 at 10:21:25 AM UTC-3, >>>>>>> Kevin Liu >>>>>>> >>>>>>>>>>>>>>>>>>>> >>>>>>> >>>>>>>>>>>>>>>>>>>> wrote: >>>>>>> >>>>>>>>>>>>>>> >>>>>> >>>>>>
