Tim Holy, what if I could tap into the well of knowledge that you are to speed up things? Can you imagine if every learner had to start without priors?
> On Aug 9, 2016, at 07:06, Tim Holy <tim.h...@gmail.com> wrote: > > I'd recommend starting by picking a very small project. For example, fix a > bug > or implement a small improvement in a package that you already find useful or > interesting. That way you'll get some guidance while making a positive > contribution; once you know more about julia, it will be easier to see your > way forward. > > Best, > --Tim > >> On Monday, August 8, 2016 8:22:01 PM CDT Kevin Liu wrote: >> I have no idea where to start and where to finish. Founders' help would be >> wonderful. >> >>> On Tuesday, August 9, 2016 at 12:19:26 AM UTC-3, Kevin Liu wrote: >>> After which I have to code Felix into Julia, a relational optimizer for >>> statistical inference with Tuffy <http://i.stanford.edu/hazy/tuffy/> >>> inside, for enterprise settings. >>> >>>> On Tuesday, August 9, 2016 at 12:07:32 AM UTC-3, Kevin Liu wrote: >>>> Can I get tips on bringing Alchemy's optimized Tuffy >>>> <http://i.stanford.edu/hazy/tuffy/> in Java to Julia while showing the >>>> best of Julia? I am going for the most correct way, even if it means >>>> coding >>>> Tuffy into C and Julia. >>>> >>>>> On Sunday, August 7, 2016 at 8:34:37 PM UTC-3, Kevin Liu wrote: >>>>> I'll try to build it, compare it, and show it to you guys. I offered to >>>>> do this as work. I am waiting to see if they will accept it. >>>>> >>>>>> On Sunday, August 7, 2016 at 6:15:50 PM UTC-3, Stefan Karpinski wrote: >>>>>> Kevin, as previously requested by Isaiah, please take this to some >>>>>> other forum or maybe start a blog. >>>>>> >>>>>>> On Sat, Aug 6, 2016 at 10:53 PM, Kevin Liu <kvt...@gmail.com> wrote: >>>>>>> Symmetry-based learning, Domingos, 2014 >>>>>>> https://www.microsoft.com/en-us/research/video/symmetry-based-learning >>>>>>> / >>>>>>> >>>>>>> Approach 2: Deep symmetry networks generalize convolutional neural >>>>>>> networks by tying parameters and pooling over an arbitrary symmetry >>>>>>> group, >>>>>>> not just the translation group. In preliminary experiments, they >>>>>>> outperformed convnets on a digit recognition task. >>>>>>> >>>>>>>> On Friday, August 5, 2016 at 4:56:45 PM UTC-3, Kevin Liu wrote: >>>>>>>> Minsky died of a cerebral hemorrhage at the age of 88.[40] >>>>>>>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-40> Ray >>>>>>>> Kurzweil <https://en.wikipedia.org/wiki/Ray_Kurzweil> says he was >>>>>>>> contacted by the cryonics organization Alcor Life Extension >>>>>>>> Foundation >>>>>>>> <https://en.wikipedia.org/wiki/Alcor_Life_Extension_Foundation> >>>>>>>> seeking >>>>>>>> Minsky's body.[41] >>>>>>>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-Kurzweil-41> >>>>>>>> Kurzweil believes that Minsky was cryonically preserved by Alcor and >>>>>>>> will be revived by 2045.[41] >>>>>>>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-Kurzweil-41> >>>>>>>> Minsky >>>>>>>> was a member of Alcor's Scientific Advisory Board >>>>>>>> <https://en.wikipedia.org/wiki/Advisory_Board>.[42] >>>>>>>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-AlcorBoard-42> >>>>>>>> In >>>>>>>> keeping with their policy of protecting privacy, Alcor will neither >>>>>>>> confirm >>>>>>>> nor deny that Alcor has cryonically preserved Minsky.[43] >>>>>>>> <https://en.wikipedia.org/wiki/Marvin_Minsky#cite_note-43> >>>>>>>> >>>>>>>> We better do a good job. >>>>>>>> >>>>>>>>> On Friday, August 5, 2016 at 4:45:42 PM UTC-3, Kevin Liu wrote: >>>>>>>>> *So, I think in the next 20 years (2003), if we can get rid of all >>>>>>>>> of the traditional approaches to artificial intelligence, like >>>>>>>>> neural nets >>>>>>>>> and genetic algorithms and rule-based systems, and just turn our >>>>>>>>> sights a >>>>>>>>> little bit higher to say, can we make a system that can use all >>>>>>>>> those >>>>>>>>> things for the right kind of problem? Some problems are good for >>>>>>>>> neural >>>>>>>>> nets; we know that others, neural nets are hopeless on them. Genetic >>>>>>>>> algorithms are great for certain things; I suspect I know what >>>>>>>>> they're bad >>>>>>>>> at, and I won't tell you. (Laughter)* - Minsky, founder of CSAIL >>>>>>>>> MIT >>>>>>>>> >>>>>>>>> *Those programmers tried to find the single best way to represent >>>>>>>>> knowledge - Only Logic protects us from paradox.* - Minsky (see >>>>>>>>> attachment from his lecture) >>>>>>>>> >>>>>>>>>> On Friday, August 5, 2016 at 8:12:03 AM UTC-3, Kevin Liu wrote: >>>>>>>>>> Markov Logic Network is being used for the continuous development >>>>>>>>>> of drugs to cure cancer at MIT's CanceRX <http://cancerx.mit.edu/>, >>>>>>>>>> on >>>>>>>>>> DARPA's largest AI project to date, Personalized Assistant that >>>>>>>>>> Learns (PAL) <https://pal.sri.com/>, progenitor of Siri. One of >>>>>>>>>> Alchemy's largest applications to date was to learn a semantic >>>>>>>>>> network >>>>>>>>>> (knowledge graph as Google calls it) from the web. >>>>>>>>>> >>>>>>>>>> Some on Probabilistic Inductive Logic Programming / Probabilistic >>>>>>>>>> Logic Programming / Statistical Relational Learning from CSAIL >>>>>>>>>> <http://people.csail.mit.edu/kersting/ecmlpkdd05_pilp/pilp_ida2005_ >>>>>>>>>> tut.pdf> (my understanding is Alchemy does PILP from entailment, >>>>>>>>>> proofs, and >>>>>>>>>> interpretation) >>>>>>>>>> >>>>>>>>>> The MIT Probabilistic Computing Project (where there is Picture, an >>>>>>>>>> extension of Julia, for computer vision; Watch the video from >>>>>>>>>> Vikash) >>>>>>>>>> <http://probcomp.csail.mit.edu/index.html> >>>>>>>>>> >>>>>>>>>> Probabilistic programming could do for Bayesian ML what Theano has >>>>>>>>>> done for neural networks. >>>>>>>>>> <http://www.inference.vc/deep-learning-is-easy/> - Ferenc Huszár >>>>>>>>>> >>>>>>>>>> Picture doesn't appear to be open-source, even though its Paper is >>>>>>>>>> available. >>>>>>>>>> >>>>>>>>>> I'm in the process of comparing the Picture Paper and Alchemy code >>>>>>>>>> and would like to have an open-source PILP from Julia that combines >>>>>>>>>> the >>>>>>>>>> best of both. >>>>>>>>>> >>>>>>>>>> On Wednesday, August 3, 2016 at 5:01:02 PM UTC-3, Christof Stocker >>>>>>>>>> >>>>>>>>>> wrote: >>>>>>>>>>> This sounds like it could be a great contribution. I shall keep a >>>>>>>>>>> curious eye on your progress >>>>>>>>>>> >>>>>>>>>>> Am Mittwoch, 3. August 2016 21:53:54 UTC+2 schrieb Kevin Liu: >>>>>>>>>>>> Thanks for the advice Cristof. I am only interested in people >>>>>>>>>>>> wanting to code it in Julia, from R by Domingos. The algo has >>>>>>>>>>>> been >>>>>>>>>>>> successfully applied in many areas, even though there are many >>>>>>>>>>>> other areas >>>>>>>>>>>> remaining. >>>>>>>>>>>> >>>>>>>>>>>> On Wed, Aug 3, 2016 at 4:45 PM, Christof Stocker < >>>>>>>>>>>> >>>>>>>>>>>> stocker....@gmail.com> wrote: >>>>>>>>>>>>> Hello Kevin, >>>>>>>>>>>>> >>>>>>>>>>>>> Enthusiasm is a good thing and you should hold on to that. But >>>>>>>>>>>>> to save yourself some headache or disappointment down the road I >>>>>>>>>>>>> advice a >>>>>>>>>>>>> level head. Nothing is really as bluntly obviously solved as it >>>>>>>>>>>>> may seems >>>>>>>>>>>>> at first glance after listening to brilliant people explain >>>>>>>>>>>>> things. A >>>>>>>>>>>>> physics professor of mine once told me that one of the (he >>>>>>>>>>>>> thinks) most >>>>>>>>>>>>> malicious factors to his past students progress where overstated >>>>>>>>>>>>> results/conclusions by other researches (such as premature >>>>>>>>>>>>> announcements >>>>>>>>>>>>> from CERN). I am no mathematician, but as far as I can judge is >>>>>>>>>>>>> the no free >>>>>>>>>>>>> lunch theorem of pure mathematical nature and not something >>>>>>>>>>>>> induced >>>>>>>>>>>>> empirically. These kind of results are not that easily to get >>>>>>>>>>>>> rid of. If >>>>>>>>>>>>> someone (especially an expert) states such a theorem will prove >>>>>>>>>>>>> wrong I >>>>>>>>>>>>> would be inclined to believe that he is not talking about >>>>>>>>>>>>> literally, but >>>>>>>>>>>>> instead is just trying to make a point about a more or less >>>>>>>>>>>>> practical >>>>>>>>>>>>> implication. >>>>>>>>>>>>> >>>>>>>>>>>>> Am Mittwoch, 3. August 2016 21:27:05 UTC+2 schrieb Kevin Liu: >>>>>>>>>>>>>> The Markov logic network represents a probability distribution >>>>>>>>>>>>>> over the states of a complex system (i.e. a cell), comprised of >>>>>>>>>>>>>> entities, >>>>>>>>>>>>>> where logic formulas encode the dependencies between them. >>>>>>>>>>>>>> >>>>>>>>>>>>>> On Wednesday, August 3, 2016 at 4:19:09 PM UTC-3, Kevin Liu >>>>>>>>>>>>>> >>>>>>>>>>>>>> wrote: >>>>>>>>>>>>>>> Alchemy is like an inductive Turing machine, to be programmed >>>>>>>>>>>>>>> to learn broadly or restrictedly. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> The logic formulas from rules through which it represents can >>>>>>>>>>>>>>> be inconsistent, incomplete, or even incorrect-- the learning >>>>>>>>>>>>>>> and >>>>>>>>>>>>>>> probabilistic reasoning will correct them. The key point is >>>>>>>>>>>>>>> that Alchemy >>>>>>>>>>>>>>> doesn't have to learn from scratch, proving Wolpert and >>>>>>>>>>>>>>> Macready's no free >>>>>>>>>>>>>>> lunch theorem wrong by performing well on a variety of classes >>>>>>>>>>>>>>> of problems, >>>>>>>>>>>>>>> not just some. >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> On Wednesday, August 3, 2016 at 4:01:15 PM UTC-3, Kevin Liu >>>>>>>>>>>>>>> >>>>>>>>>>>>>>> wrote: >>>>>>>>>>>>>>>> Hello Community, >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> I'm in the last pages of Pedro Domingos' book, the Master >>>>>>>>>>>>>>>> Algo, one of two recommended by Bill Gates to learn about AI. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> From the book, I understand all learners have to represent, >>>>>>>>>>>>>>>> evaluate, and optimize. There are many types of learners that >>>>>>>>>>>>>>>> do this. What >>>>>>>>>>>>>>>> Domingos does is generalize these three parts, (1) using >>>>>>>>>>>>>>>> Markov Logic >>>>>>>>>>>>>>>> Network to represent, (2) posterior probability to evaluate, >>>>>>>>>>>>>>>> and (3) >>>>>>>>>>>>>>>> genetic search with gradient descent to optimize. The >>>>>>>>>>>>>>>> posterior can be >>>>>>>>>>>>>>>> replaced for another accuracy measure when it is easier, as >>>>>>>>>>>>>>>> genetic search >>>>>>>>>>>>>>>> replaced by hill climbing. Where there are 15 popular options >>>>>>>>>>>>>>>> for >>>>>>>>>>>>>>>> representing, evaluating, and optimizing, Domingos >>>>>>>>>>>>>>>> generalized them into >>>>>>>>>>>>>>>> three options. The idea is to have one unified learner for >>>>>>>>>>>>>>>> any application. >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> There is code already done in R >>>>>>>>>>>>>>>> https://alchemy.cs.washington.edu/. My question: anybody in >>>>>>>>>>>>>>>> the community vested in coding it into Julia? >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>> Thanks. Kevin >>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> On Friday, June 3, 2016 at 3:44:09 PM UTC-3, Kevin Liu wrote: >>>>>>>>>>>>>>>>> https://github.com/tbreloff/OnlineAI.jl/issues/5 >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> On Friday, June 3, 2016 at 11:17:28 AM UTC-3, Kevin Liu >>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>> wrote: >>>>>>>>>>>>>>>>>> I plan to write Julia for the rest of me life... given it >>>>>>>>>>>>>>>>>> remains suitable. I am still reading all of Colah's >>>>>>>>>>>>>>>>>> material on nets. I ran >>>>>>>>>>>>>>>>>> Mocha.jl a couple weeks ago and was very happy to see it >>>>>>>>>>>>>>>>>> work. Thanks for >>>>>>>>>>>>>>>>>> jumping in and telling me about OnlineAI.jl, I will look >>>>>>>>>>>>>>>>>> into it once I am >>>>>>>>>>>>>>>>>> ready. From a quick look, perhaps I could help and learn by >>>>>>>>>>>>>>>>>> building a very >>>>>>>>>>>>>>>>>> clear documentation of it. Would really like to see Julia a >>>>>>>>>>>>>>>>>> leap ahead of >>>>>>>>>>>>>>>>>> other languages, and plan to contribute heavily to it, but >>>>>>>>>>>>>>>>>> at the moment am >>>>>>>>>>>>>>>>>> still getting introduced to CS, programming, and nets at >>>>>>>>>>>>>>>>>> the basic level. >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> On Friday, June 3, 2016 at 10:48:15 AM UTC-3, Tom Breloff >>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>> wrote: >>>>>>>>>>>>>>>>>>> Kevin: computers that program themselves is a concept >>>>>>>>>>>>>>>>>>> which is much closer to reality than most would believe, >>>>>>>>>>>>>>>>>>> but julia-users >>>>>>>>>>>>>>>>>>> isn't really the best place for this speculation. If >>>>>>>>>>>>>>>>>>> you're actually >>>>>>>>>>>>>>>>>>> interested in writing code, I'm happy to discuss in >>>>>>>>>>>>>>>>>>> OnlineAI.jl. I was >>>>>>>>>>>>>>>>>>> thinking about how we might tackle code generation using a >>>>>>>>>>>>>>>>>>> neural framework >>>>>>>>>>>>>>>>>>> I'm working on. >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> On Friday, June 3, 2016, Kevin Liu <kvt...@gmail.com> >>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>> wrote: >>>>>>>>>>>>>>>>>>>> If Andrew Ng who cited Gates, and Gates who cited >>>>>>>>>>>>>>>>>>>> Domingos (who did not lecture at Google with a TensorFlow >>>>>>>>>>>>>>>>>>>> question in the >>>>>>>>>>>>>>>>>>>> end), were unsuccessful penny traders, Julia was a >>>>>>>>>>>>>>>>>>>> language for web design, >>>>>>>>>>>>>>>>>>>> and the tribes in the video didn't actually solve >>>>>>>>>>>>>>>>>>>> problems, perhaps this >>>>>>>>>>>>>>>>>>>> would be a wildly off-topic, speculative discussion. But >>>>>>>>>>>>>>>>>>>> these statements >>>>>>>>>>>>>>>>>>>> couldn't be farther from the truth. In fact, if I had >>>>>>>>>>>>>>>>>>>> known about this >>>>>>>>>>>>>>>>>>>> video some months ago I would've understood better on how >>>>>>>>>>>>>>>>>>>> to solve a >>>>>>>>>>>>>>>>>>>> problem I was working on. >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> For the founders of Julia: I understand your tribe is >>>>>>>>>>>>>>>>>>>> mainly CS. This master algorithm, as you are aware, would >>>>>>>>>>>>>>>>>>>> require >>>>>>>>>>>>>>>>>>>> collaboration with other tribes. Just citing the obvious. >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> On Friday, June 3, 2016 at 10:21:25 AM UTC-3, Kevin Liu >>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>> wrote: >>>>>>>>>>>>>>>>>>>>> There could be parts missing as Domingos mentions, but >>>>>>>>>>>>>>>>>>>>> induction, backpropagation, genetic programming, >>>>>>>>>>>>>>>>>>>>> probabilistic inference, >>>>>>>>>>>>>>>>>>>>> and SVMs working together-- what's speculative about the >>>>>>>>>>>>>>>>>>>>> improved versions >>>>>>>>>>>>>>>>>>>>> of these? >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> Julia was made for AI. Isn't it time for a consolidated >>>>>>>>>>>>>>>>>>>>> view on how to reach it? >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> On Thursday, June 2, 2016 at 11:20:35 PM UTC-3, Isaiah >>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>> wrote: >>>>>>>>>>>>>>>>>>>>>> This is not a forum for wildly off-topic, speculative >>>>>>>>>>>>>>>>>>>>>> discussion. >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> Take this to Reddit, Hacker News, etc. >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> On Thu, Jun 2, 2016 at 10:01 PM, Kevin Liu < >>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>> kvt...@gmail.com> wrote: >>>>>>>>>>>>>>>>>>>>>>> I am wondering how Julia fits in with the unified >>>>>>>>>>>>>>>>>>>>>>> tribes >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> mashable.com/2016/06/01/bill-gates-ai-code-conference/ >>>>>>>>>>>>>>>>>>>>>>> #8VmBFjIiYOqJ >>>>>>>>>>>>>>>>>>>>>>> >>>>>>>>>>>>>>>>>>>>>>> https://www.youtube.com/watch?v=B8J4uefCQMc > >