I'd recommend starting by picking a very small project. For example, fix a bug 
or implement a small improvement in a package that you already find useful or 
interesting. That way you'll get some guidance while making a positive 
contribution; once you know more about julia, it will be easier to see your 
way forward.

Best,
--Tim

On Monday, August 8, 2016 8:22:01 PM CDT Kevin Liu wrote:
> I have no idea where to start and where to finish. Founders' help would be
> wonderful.
> 
> On Tuesday, August 9, 2016 at 12:19:26 AM UTC-3, Kevin Liu wrote:
> > After which I have to code Felix into Julia, a relational optimizer for
> > statistical inference with Tuffy <http://i.stanford.edu/hazy/tuffy/>
> > inside, for enterprise settings.
> > 
> > On Tuesday, August 9, 2016 at 12:07:32 AM UTC-3, Kevin Liu wrote:
> >> Can I get tips on bringing Alchemy's optimized Tuffy
> >> <http://i.stanford.edu/hazy/tuffy/> in Java to Julia while showing the
> >> best of Julia? I am going for the most correct way, even if it means
> >> coding
> >> Tuffy into C and Julia.
> >> 
> >> On Sunday, August 7, 2016 at 8:34:37 PM UTC-3, Kevin Liu wrote:
> >>> I'll try to build it, compare it, and show it to you guys. I offered to
> >>> do this as work. I am waiting to see if they will accept it.
> >>> 
> >>> On Sunday, August 7, 2016 at 6:15:50 PM UTC-3, Stefan Karpinski wrote:
> >>>> Kevin, as previously requested by Isaiah, please take this to some
> >>>> other forum or maybe start a blog.
> >>>> 
> >>>> On Sat, Aug 6, 2016 at 10:53 PM, Kevin Liu <[email protected]> wrote:
> >>>>> Symmetry-based learning, Domingos, 2014
> >>>>> https://www.microsoft.com/en-us/research/video/symmetry-based-learning
> >>>>> /
> >>>>> 
> >>>>> Approach 2: Deep symmetry networks generalize convolutional neural
> >>>>> networks by tying parameters and pooling over an arbitrary symmetry
> >>>>> group,
> >>>>> not just the translation group. In preliminary experiments, they
> >>>>> outperformed convnets on a digit recognition task.
> >>>>> 
> >>>>> On Friday, August 5, 2016 at 4:56:45 PM UTC-3, Kevin Liu wrote:
> >>>>>> Minsky died of a cerebral hemorrhage at the age of 88.[40]
> >>>>>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-40> Ray
> >>>>>> Kurzweil <https://en.wikipedia.org/wiki/Ray_Kurzweil> says he was
> >>>>>> contacted by the cryonics organization Alcor Life Extension
> >>>>>> Foundation
> >>>>>> <https://en.wikipedia.org/wiki/Alcor_Life_Extension_Foundation>
> >>>>>> seeking
> >>>>>> Minsky's body.[41]
> >>>>>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-Kurzweil-41>
> >>>>>> Kurzweil believes that Minsky was cryonically preserved by Alcor and
> >>>>>> will be revived by 2045.[41]
> >>>>>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-Kurzweil-41>
> >>>>>> Minsky
> >>>>>> was a member of Alcor's Scientific Advisory Board
> >>>>>> <https://en.wikipedia.org/wiki/Advisory_Board>.[42]
> >>>>>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-AlcorBoard-42>
> >>>>>> In
> >>>>>> keeping with their policy of protecting privacy, Alcor will neither
> >>>>>> confirm
> >>>>>> nor deny that Alcor has cryonically preserved Minsky.[43]
> >>>>>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-43>
> >>>>>> 
> >>>>>> We better do a good job.
> >>>>>> 
> >>>>>> On Friday, August 5, 2016 at 4:45:42 PM UTC-3, Kevin Liu wrote:
> >>>>>>> *So, I think in the next 20 years (2003), if we can get rid of all
> >>>>>>> of the traditional approaches to artificial intelligence, like
> >>>>>>> neural nets
> >>>>>>> and genetic algorithms and rule-based systems, and just turn our
> >>>>>>> sights a
> >>>>>>> little bit higher to say, can we make a system that can use all
> >>>>>>> those
> >>>>>>> things for the right kind of problem? Some problems are good for
> >>>>>>> neural
> >>>>>>> nets; we know that others, neural nets are hopeless on them. Genetic
> >>>>>>> algorithms are great for certain things; I suspect I know what
> >>>>>>> they're bad
> >>>>>>> at, and I won't tell you. (Laughter)*  - Minsky, founder of CSAIL
> >>>>>>> MIT
> >>>>>>> 
> >>>>>>> *Those programmers tried to find the single best way to represent
> >>>>>>> knowledge - Only Logic protects us from paradox.* - Minsky (see
> >>>>>>> attachment from his lecture)
> >>>>>>> 
> >>>>>>> On Friday, August 5, 2016 at 8:12:03 AM UTC-3, Kevin Liu wrote:
> >>>>>>>> Markov Logic Network is being used for the continuous development
> >>>>>>>> of drugs to cure cancer at MIT's CanceRX <http://cancerx.mit.edu/>,
> >>>>>>>> on
> >>>>>>>> DARPA's largest AI project to date, Personalized Assistant that
> >>>>>>>> Learns (PAL) <https://pal.sri.com/>, progenitor of Siri. One of
> >>>>>>>> Alchemy's largest applications to date was to learn a semantic
> >>>>>>>> network
> >>>>>>>> (knowledge graph as Google calls it) from the web.
> >>>>>>>> 
> >>>>>>>> Some on Probabilistic Inductive Logic Programming / Probabilistic
> >>>>>>>> Logic Programming / Statistical Relational Learning from CSAIL
> >>>>>>>> <http://people.csail.mit.edu/kersting/ecmlpkdd05_pilp/pilp_ida2005_
> >>>>>>>> tut.pdf> (my understanding is Alchemy does PILP from entailment,
> >>>>>>>> proofs, and
> >>>>>>>> interpretation)
> >>>>>>>> 
> >>>>>>>> The MIT Probabilistic Computing Project (where there is Picture, an
> >>>>>>>> extension of Julia, for computer vision; Watch the video from
> >>>>>>>> Vikash)
> >>>>>>>> <http://probcomp.csail.mit.edu/index.html>
> >>>>>>>> 
> >>>>>>>> Probabilistic programming could do for Bayesian ML what Theano has
> >>>>>>>> done for neural networks.
> >>>>>>>> <http://www.inference.vc/deep-learning-is-easy/> - Ferenc Huszár
> >>>>>>>> 
> >>>>>>>> Picture doesn't appear to be open-source, even though its Paper is
> >>>>>>>> available.
> >>>>>>>> 
> >>>>>>>> I'm in the process of comparing the Picture Paper and Alchemy code
> >>>>>>>> and would like to have an open-source PILP from Julia that combines
> >>>>>>>> the
> >>>>>>>> best of both.
> >>>>>>>> 
> >>>>>>>> On Wednesday, August 3, 2016 at 5:01:02 PM UTC-3, Christof Stocker
> >>>>>>>> 
> >>>>>>>> wrote:
> >>>>>>>>> This sounds like it could be a great contribution. I shall keep a
> >>>>>>>>> curious eye on your progress
> >>>>>>>>> 
> >>>>>>>>> Am Mittwoch, 3. August 2016 21:53:54 UTC+2 schrieb Kevin Liu:
> >>>>>>>>>> Thanks for the advice Cristof. I am only interested in people
> >>>>>>>>>> wanting to code it in Julia, from R by Domingos. The algo has
> >>>>>>>>>> been
> >>>>>>>>>> successfully applied in many areas, even though there are many
> >>>>>>>>>> other areas
> >>>>>>>>>> remaining.
> >>>>>>>>>> 
> >>>>>>>>>> On Wed, Aug 3, 2016 at 4:45 PM, Christof Stocker <
> >>>>>>>>>> 
> >>>>>>>>>> [email protected]> wrote:
> >>>>>>>>>>> Hello Kevin,
> >>>>>>>>>>> 
> >>>>>>>>>>> Enthusiasm is a good thing and you should hold on to that. But
> >>>>>>>>>>> to save yourself some headache or disappointment down the road I
> >>>>>>>>>>> advice a
> >>>>>>>>>>> level head. Nothing is really as bluntly obviously solved as it
> >>>>>>>>>>> may seems
> >>>>>>>>>>> at first glance after listening to brilliant people explain
> >>>>>>>>>>> things. A
> >>>>>>>>>>> physics professor of mine once told me that one of the (he
> >>>>>>>>>>> thinks) most
> >>>>>>>>>>> malicious factors to his past students progress where overstated
> >>>>>>>>>>> results/conclusions by other researches (such as premature
> >>>>>>>>>>> announcements
> >>>>>>>>>>> from CERN). I am no mathematician, but as far as I can judge is
> >>>>>>>>>>> the no free
> >>>>>>>>>>> lunch theorem of pure mathematical nature and not something
> >>>>>>>>>>> induced
> >>>>>>>>>>> empirically. These kind of results are not that easily to get
> >>>>>>>>>>> rid of. If
> >>>>>>>>>>> someone (especially an expert) states such a theorem will prove
> >>>>>>>>>>> wrong I
> >>>>>>>>>>> would be inclined to believe that he is not talking about
> >>>>>>>>>>> literally, but
> >>>>>>>>>>> instead is just trying to make a point about a more or less
> >>>>>>>>>>> practical
> >>>>>>>>>>> implication.
> >>>>>>>>>>> 
> >>>>>>>>>>> Am Mittwoch, 3. August 2016 21:27:05 UTC+2 schrieb Kevin Liu:
> >>>>>>>>>>>> The Markov logic network represents a probability distribution
> >>>>>>>>>>>> over the states of a complex system (i.e. a cell), comprised of
> >>>>>>>>>>>> entities,
> >>>>>>>>>>>> where logic formulas encode the dependencies between them.
> >>>>>>>>>>>> 
> >>>>>>>>>>>> On Wednesday, August 3, 2016 at 4:19:09 PM UTC-3, Kevin Liu
> >>>>>>>>>>>> 
> >>>>>>>>>>>> wrote:
> >>>>>>>>>>>>> Alchemy is like an inductive Turing machine, to be programmed
> >>>>>>>>>>>>> to learn broadly or restrictedly.
> >>>>>>>>>>>>> 
> >>>>>>>>>>>>> The logic formulas from rules through which it represents can
> >>>>>>>>>>>>> be inconsistent, incomplete, or even incorrect-- the learning
> >>>>>>>>>>>>> and
> >>>>>>>>>>>>> probabilistic reasoning will correct them. The key point is
> >>>>>>>>>>>>> that Alchemy
> >>>>>>>>>>>>> doesn't have to learn from scratch, proving Wolpert and
> >>>>>>>>>>>>> Macready's no free
> >>>>>>>>>>>>> lunch theorem wrong by performing well on a variety of classes
> >>>>>>>>>>>>> of problems,
> >>>>>>>>>>>>> not just some.
> >>>>>>>>>>>>> 
> >>>>>>>>>>>>> On Wednesday, August 3, 2016 at 4:01:15 PM UTC-3, Kevin Liu
> >>>>>>>>>>>>> 
> >>>>>>>>>>>>> wrote:
> >>>>>>>>>>>>>> Hello Community,
> >>>>>>>>>>>>>> 
> >>>>>>>>>>>>>> I'm in the last pages of Pedro Domingos' book, the Master
> >>>>>>>>>>>>>> Algo, one of two recommended by Bill Gates to learn about AI.
> >>>>>>>>>>>>>> 
> >>>>>>>>>>>>>> From the book, I understand all learners have to represent,
> >>>>>>>>>>>>>> evaluate, and optimize. There are many types of learners that
> >>>>>>>>>>>>>> do this. What
> >>>>>>>>>>>>>> Domingos does is generalize these three parts, (1) using
> >>>>>>>>>>>>>> Markov Logic
> >>>>>>>>>>>>>> Network to represent, (2) posterior probability to evaluate,
> >>>>>>>>>>>>>> and (3)
> >>>>>>>>>>>>>> genetic search with gradient descent to optimize. The
> >>>>>>>>>>>>>> posterior can be
> >>>>>>>>>>>>>> replaced for another accuracy measure when it is easier, as
> >>>>>>>>>>>>>> genetic search
> >>>>>>>>>>>>>> replaced by hill climbing. Where there are 15 popular options
> >>>>>>>>>>>>>> for
> >>>>>>>>>>>>>> representing, evaluating, and optimizing, Domingos
> >>>>>>>>>>>>>> generalized them into
> >>>>>>>>>>>>>> three options. The idea is to have one unified learner for
> >>>>>>>>>>>>>> any application.
> >>>>>>>>>>>>>> 
> >>>>>>>>>>>>>> There is code already done in R
> >>>>>>>>>>>>>> https://alchemy.cs.washington.edu/. My question: anybody in
> >>>>>>>>>>>>>> the community vested in coding it into Julia?
> >>>>>>>>>>>>>> 
> >>>>>>>>>>>>>> Thanks. Kevin
> >>>>>>>>>>>>>> 
> >>>>>>>>>>>>>> On Friday, June 3, 2016 at 3:44:09 PM UTC-3, Kevin Liu wrote:
> >>>>>>>>>>>>>>> https://github.com/tbreloff/OnlineAI.jl/issues/5
> >>>>>>>>>>>>>>> 
> >>>>>>>>>>>>>>> On Friday, June 3, 2016 at 11:17:28 AM UTC-3, Kevin Liu
> >>>>>>>>>>>>>>> 
> >>>>>>>>>>>>>>> wrote:
> >>>>>>>>>>>>>>>> I plan to write Julia for the rest of me life... given it
> >>>>>>>>>>>>>>>> remains suitable. I am still reading all of Colah's
> >>>>>>>>>>>>>>>> material on nets. I ran
> >>>>>>>>>>>>>>>> Mocha.jl a couple weeks ago and was very happy to see it
> >>>>>>>>>>>>>>>> work. Thanks for
> >>>>>>>>>>>>>>>> jumping in and telling me about OnlineAI.jl, I will look
> >>>>>>>>>>>>>>>> into it once I am
> >>>>>>>>>>>>>>>> ready. From a quick look, perhaps I could help and learn by
> >>>>>>>>>>>>>>>> building a very
> >>>>>>>>>>>>>>>> clear documentation of it. Would really like to see Julia a
> >>>>>>>>>>>>>>>> leap ahead of
> >>>>>>>>>>>>>>>> other languages, and plan to contribute heavily to it, but
> >>>>>>>>>>>>>>>> at the moment am
> >>>>>>>>>>>>>>>> still getting introduced to CS, programming, and nets at
> >>>>>>>>>>>>>>>> the basic level.
> >>>>>>>>>>>>>>>> 
> >>>>>>>>>>>>>>>> On Friday, June 3, 2016 at 10:48:15 AM UTC-3, Tom Breloff
> >>>>>>>>>>>>>>>> 
> >>>>>>>>>>>>>>>> wrote:
> >>>>>>>>>>>>>>>>> Kevin: computers that program themselves is a concept
> >>>>>>>>>>>>>>>>> which is much closer to reality than most would believe,
> >>>>>>>>>>>>>>>>> but julia-users
> >>>>>>>>>>>>>>>>> isn't really the best place for this speculation. If
> >>>>>>>>>>>>>>>>> you're actually
> >>>>>>>>>>>>>>>>> interested in writing code, I'm happy to discuss in
> >>>>>>>>>>>>>>>>> OnlineAI.jl. I was
> >>>>>>>>>>>>>>>>> thinking about how we might tackle code generation using a
> >>>>>>>>>>>>>>>>> neural framework
> >>>>>>>>>>>>>>>>> I'm working on.
> >>>>>>>>>>>>>>>>> 
> >>>>>>>>>>>>>>>>> On Friday, June 3, 2016, Kevin Liu <[email protected]>
> >>>>>>>>>>>>>>>>> 
> >>>>>>>>>>>>>>>>> wrote:
> >>>>>>>>>>>>>>>>>> If Andrew Ng who cited Gates, and Gates who cited
> >>>>>>>>>>>>>>>>>> Domingos (who did not lecture at Google with a TensorFlow
> >>>>>>>>>>>>>>>>>> question in the
> >>>>>>>>>>>>>>>>>> end), were unsuccessful penny traders, Julia was a
> >>>>>>>>>>>>>>>>>> language for web design,
> >>>>>>>>>>>>>>>>>> and the tribes in the video didn't actually solve
> >>>>>>>>>>>>>>>>>> problems, perhaps this
> >>>>>>>>>>>>>>>>>> would be a wildly off-topic, speculative discussion. But
> >>>>>>>>>>>>>>>>>> these statements
> >>>>>>>>>>>>>>>>>> couldn't be farther from the truth. In fact, if I had
> >>>>>>>>>>>>>>>>>> known about this
> >>>>>>>>>>>>>>>>>> video some months ago I would've understood better on how
> >>>>>>>>>>>>>>>>>> to solve a
> >>>>>>>>>>>>>>>>>> problem I was working on.
> >>>>>>>>>>>>>>>>>> 
> >>>>>>>>>>>>>>>>>> For the founders of Julia: I understand your tribe is
> >>>>>>>>>>>>>>>>>> mainly CS. This master algorithm, as you are aware, would
> >>>>>>>>>>>>>>>>>> require
> >>>>>>>>>>>>>>>>>> collaboration with other tribes. Just citing the obvious.
> >>>>>>>>>>>>>>>>>> 
> >>>>>>>>>>>>>>>>>> On Friday, June 3, 2016 at 10:21:25 AM UTC-3, Kevin Liu
> >>>>>>>>>>>>>>>>>> 
> >>>>>>>>>>>>>>>>>> wrote:
> >>>>>>>>>>>>>>>>>>> There could be parts missing as Domingos mentions, but
> >>>>>>>>>>>>>>>>>>> induction, backpropagation, genetic programming,
> >>>>>>>>>>>>>>>>>>> probabilistic inference,
> >>>>>>>>>>>>>>>>>>> and SVMs working together-- what's speculative about the
> >>>>>>>>>>>>>>>>>>> improved versions
> >>>>>>>>>>>>>>>>>>> of these?
> >>>>>>>>>>>>>>>>>>> 
> >>>>>>>>>>>>>>>>>>> Julia was made for AI. Isn't it time for a consolidated
> >>>>>>>>>>>>>>>>>>> view on how to reach it?
> >>>>>>>>>>>>>>>>>>> 
> >>>>>>>>>>>>>>>>>>> On Thursday, June 2, 2016 at 11:20:35 PM UTC-3, Isaiah
> >>>>>>>>>>>>>>>>>>> 
> >>>>>>>>>>>>>>>>>>> wrote:
> >>>>>>>>>>>>>>>>>>>> This is not a forum for wildly off-topic, speculative
> >>>>>>>>>>>>>>>>>>>> discussion.
> >>>>>>>>>>>>>>>>>>>> 
> >>>>>>>>>>>>>>>>>>>> Take this to Reddit, Hacker News, etc.
> >>>>>>>>>>>>>>>>>>>> 
> >>>>>>>>>>>>>>>>>>>> 
> >>>>>>>>>>>>>>>>>>>> On Thu, Jun 2, 2016 at 10:01 PM, Kevin Liu <
> >>>>>>>>>>>>>>>>>>>> 
> >>>>>>>>>>>>>>>>>>>> [email protected]> wrote:
> >>>>>>>>>>>>>>>>>>>>> I am wondering how Julia fits in with the unified
> >>>>>>>>>>>>>>>>>>>>> tribes
> >>>>>>>>>>>>>>>>>>>>> 
> >>>>>>>>>>>>>>>>>>>>> 
> >>>>>>>>>>>>>>>>>>>>> mashable.com/2016/06/01/bill-gates-ai-code-conference/
> >>>>>>>>>>>>>>>>>>>>> #8VmBFjIiYOqJ
> >>>>>>>>>>>>>>>>>>>>> 
> >>>>>>>>>>>>>>>>>>>>> https://www.youtube.com/watch?v=B8J4uefCQMc


Reply via email to