Another important cool thing I think is worth noting: he added the 
possibility of weights to rules (attachment). Each line is equivalent to a 
desired conclusion. 

On Wednesday, August 3, 2016 at 4:27:05 PM UTC-3, Kevin Liu wrote:
>
> The Markov logic network represents a probability distribution over the 
> states of a complex system (i.e. a cell), comprised of entities, where 
> logic formulas encode the dependencies between them. 
>
> On Wednesday, August 3, 2016 at 4:19:09 PM UTC-3, Kevin Liu wrote:
>>
>> Alchemy is like an inductive Turing machine, to be programmed to learn 
>> broadly or restrictedly.
>>
>> The logic formulas from rules through which it represents can be 
>> inconsistent, incomplete, or even incorrect-- the learning and 
>> probabilistic reasoning will correct them. The key point is that Alchemy 
>> doesn't have to learn from scratch, proving Wolpert and Macready's no free 
>> lunch theorem wrong by performing well on a variety of classes of problems, 
>> not just some.
>>
>> On Wednesday, August 3, 2016 at 4:01:15 PM UTC-3, Kevin Liu wrote:
>>>
>>> Hello Community, 
>>>
>>> I'm in the last pages of Pedro Domingos' book, the Master Algo, one of 
>>> two recommended by Bill Gates to learn about AI. 
>>>
>>> From the book, I understand all learners have to represent, evaluate, 
>>> and optimize. There are many types of learners that do this. What Domingos 
>>> does is generalize these three parts, (1) using Markov Logic Network to 
>>> represent, (2) posterior probability to evaluate, and (3) genetic search 
>>> with gradient descent to optimize. The posterior can be replaced for 
>>> another accuracy measure when it is easier, as genetic search replaced by 
>>> hill climbing. Where there are 15 popular options for representing, 
>>> evaluating, and optimizing, Domingos generalized them into three options. 
>>> The idea is to have one unified learner for any application. 
>>>
>>> There is code already done in R https://alchemy.cs.washington.edu/. My 
>>> question: anybody in the community vested in coding it into Julia?
>>>
>>> Thanks. Kevin
>>>
>>> On Friday, June 3, 2016 at 3:44:09 PM UTC-3, Kevin Liu wrote:
>>>>
>>>> https://github.com/tbreloff/OnlineAI.jl/issues/5
>>>>
>>>> On Friday, June 3, 2016 at 11:17:28 AM UTC-3, Kevin Liu wrote:
>>>>>
>>>>> I plan to write Julia for the rest of me life... given it remains 
>>>>> suitable. I am still reading all of Colah's material on nets. I ran 
>>>>> Mocha.jl a couple weeks ago and was very happy to see it work. Thanks for 
>>>>> jumping in and telling me about OnlineAI.jl, I will look into it once I 
>>>>> am 
>>>>> ready. From a quick look, perhaps I could help and learn by building a 
>>>>> very 
>>>>> clear documentation of it. Would really like to see Julia a leap ahead of 
>>>>> other languages, and plan to contribute heavily to it, but at the moment 
>>>>> am 
>>>>> still getting introduced to CS, programming, and nets at the basic level. 
>>>>>
>>>>> On Friday, June 3, 2016 at 10:48:15 AM UTC-3, Tom Breloff wrote:
>>>>>>
>>>>>> Kevin: computers that program themselves is a concept which is much 
>>>>>> closer to reality than most would believe, but julia-users isn't really 
>>>>>> the 
>>>>>> best place for this speculation. If you're actually interested in 
>>>>>> writing 
>>>>>> code, I'm happy to discuss in OnlineAI.jl. I was thinking about how we 
>>>>>> might tackle code generation using a neural framework I'm working on. 
>>>>>>
>>>>>> On Friday, June 3, 2016, Kevin Liu <kvt...@gmail.com> wrote:
>>>>>>
>>>>>>> If Andrew Ng who cited Gates, and Gates who cited Domingos (who did 
>>>>>>> not lecture at Google with a TensorFlow question in the end), were 
>>>>>>> unsuccessful penny traders, Julia was a language for web design, and 
>>>>>>> the 
>>>>>>> tribes in the video didn't actually solve problems, perhaps this would 
>>>>>>> be a 
>>>>>>> wildly off-topic, speculative discussion. But these statements couldn't 
>>>>>>> be 
>>>>>>> farther from the truth. In fact, if I had known about this video some 
>>>>>>> months ago I would've understood better on how to solve a problem I was 
>>>>>>> working on.  
>>>>>>>
>>>>>>> For the founders of Julia: I understand your tribe is mainly CS. 
>>>>>>> This master algorithm, as you are aware, would require collaboration 
>>>>>>> with 
>>>>>>> other tribes. Just citing the obvious. 
>>>>>>>
>>>>>>> On Friday, June 3, 2016 at 10:21:25 AM UTC-3, Kevin Liu wrote:
>>>>>>>>
>>>>>>>> There could be parts missing as Domingos mentions, but induction, 
>>>>>>>> backpropagation, genetic programming, probabilistic inference, and 
>>>>>>>> SVMs 
>>>>>>>> working together-- what's speculative about the improved versions of 
>>>>>>>> these? 
>>>>>>>>
>>>>>>>> Julia was made for AI. Isn't it time for a consolidated view on how 
>>>>>>>> to reach it? 
>>>>>>>>
>>>>>>>> On Thursday, June 2, 2016 at 11:20:35 PM UTC-3, Isaiah wrote:
>>>>>>>>>
>>>>>>>>> This is not a forum for wildly off-topic, speculative discussion.
>>>>>>>>>
>>>>>>>>> Take this to Reddit, Hacker News, etc.
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Thu, Jun 2, 2016 at 10:01 PM, Kevin Liu <kvt...@gmail.com> 
>>>>>>>>> wrote:
>>>>>>>>>
>>>>>>>>>> I am wondering how Julia fits in with the unified tribes
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> mashable.com/2016/06/01/bill-gates-ai-code-conference/#8VmBFjIiYOqJ
>>>>>>>>>>
>>>>>>>>>> https://www.youtube.com/watch?v=B8J4uefCQMc
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>

Reply via email to