I sent a message to Anton saying "have you tried skipping the MST step
entirely?" and "does the quality of your results degrade if you do?" I kind
of expect the reply to be "huh, what do you mean by that?" but sure, that's
off-topic for this list.  --linas

p.s. re computational biology; do not underestimate the complexity of
biology. People have actually built multi-million-$$ protein-folding
hardware, and protein-folding@home is still a thing. You may be able to use
machine-learning techniques to tease out metabolic pathways, discern the
structure in them (I currently believe they're not unlike natural language
in structure) but there's a terrible combinatorial explosion in biochem
that should not be taken lightly.

--linas

On Sat, Feb 9, 2019 at 8:29 PM Ben Goertzel <b...@goertzel.org> wrote:

> Current MST step is suboptimal, but there is value in extracting a
> parse tree with no-links-crossing from the tangly graph of dependency
> links interconnecting the words in a sentence....   As you now this
> sort of pruning is key to how language works ... syntax "rules" exist
> to enable this sort of pruning, thus allowing us to pinpoint exact
> sets of semantic relationships within the broader tangly web of
> semantic relationships that our mind identifies between any set of
> concepts or objects or relationships...
>
> However, this is an in-depth technical discussion probably best not
> carried out on this general-purpose list...
>
> On Sun, Feb 10, 2019 at 3:58 AM Linas Vepstas <linasveps...@gmail.com>
> wrote:
> >
> >
> >
> > On Sat, Feb 9, 2019 at 4:22 AM Ben Goertzel <b...@goertzel.org> wrote:
> >>
> >>
> >> We are now playing with hybridizing these symbolic-ish grammar
> >> induction methods with neural net language models, basically using the
> >> predictive models produced by models in the BERT lineage (but more
> >> sophisticated than vanilla BERT) in place of simple mutual information
> >> values to produce more broadly-context-sensitive parse choices in
> >> Linas's MST parser...
> >
> >
> > This last sentence suggests that the near-total confusion about MST
> continues to persist in the team. I keep telling them to collect the
> statistics, and then discard the MST parse **immediately**. Trying to
> "improve" MST is a total waste of time.
> >
> > Seriously: Instead, try skipping the MST step entirely.  Just do not
> even do it, AT ALL. Rip it out. It is NOT a step that the algorithm even
> needs.  I'll bet you that if you skip the MST step completely, the quality
> of your results will be more-or-less unchanged.  The results might even get
> better!
> >
> > If your results don't change, by skipping MST, or if your results get
> better, by skipping MST, then that should be a clear indicator that trying
> to "improve" MST is a waste of time!
> >
> > -- Linas
> >
> > --
> > cassette tapes - analog TV - film cameras - you
> > Artificial General Intelligence List / AGI / see discussions +
> participants + delivery options Permalink
> 
> --
> Ben Goertzel, PhD
> http://goertzel.org
> 
> "The dewdrop world / Is the dewdrop world / And yet, and yet …" --
> Kobayashi Issa


-- 
cassette tapes - analog TV - film cameras - you

------------------------------------------
Artificial General Intelligence List: AGI
Permalink: 
https://agi.topicbox.com/groups/agi/Ta6fce6a7b640886a-M9a77ba22548daeb0cb7f9c62
Delivery options: https://agi.topicbox.com/groups/agi/subscription

Reply via email to