Then there's best-first search, B*, C*, constraint-directed search, etc.
And these are just classical search methods.

Feank

Frank Wimberly
Phone (505) 670-9918

On Aug 8, 2017 7:20 PM, "Marcus Daniels" <mar...@snoutfarm.com> wrote:

> "But one problem is that breadth-first and depth-first search are just
> fast ways to find answers."
>
>
> Just _not_ -- general but not efficient.   [My dog was demanding
> attention! ]
> ------------------------------
> *From:* Friam <friam-boun...@redfish.com> on behalf of Marcus Daniels <
> mar...@snoutfarm.com>
> *Sent:* Tuesday, August 8, 2017 6:43:40 PM
> *To:* The Friday Morning Applied Complexity Coffee Group; glen ☣
> *Subject:* Re: [FRIAM] Future of humans and artificial intelligence
>
>
> Grant writes:
>
>
> "On the other hand... evolution *is* stochastic. (You actually did not
> disagree with me on that. You only said that the reason I was right was
> another one.) "
>
>
> I think of logic programming systems as a traditional tool of AI research
> (e.g. Prolog, now Curry, similar capabilities implemented in Lisp) from the
> age before the AI winter.  These systems provide a very flexible way to
> pose constraint problems.  But one problem is that breadth-first and
> depth-first search are just fast ways to find answers.  Recent work seems
> to have shifted to SMT solvers and specialized constraint solving
> algorithms, but these have somewhat less expressiveness as programming
> languages.  Meanwhile, machine learning has come on the scene in a big way
> and tasks traditionally associated with old-school AI, like natural
> language processing, are now matched or even dominated using neural nets
> (LSTM).  I find the range of capabilities provided by groups like
> nlp.stanford.edu really impressive -- there examples of both approaches
> (logic programming and machine learning) and then don't need to be mutually
> exclusive.
>
>
> Quantum annealing is one area where the two may increasingly come together
> by using physical phenomena to accelerate the rate at which high
> dimensional discrete systems can be solved, without relying on fragile or
> domain-specific heuristics.
>
>
> I often use evolutionary algorithms for hard optimization problems.
> Genetic algorithms, for example, are robust to  noise (or if you like
> ambiguity) in fitness functions, and they are trivial to parallelize.
>
>
> Marcus
> ------------------------------
> *From:* Friam <friam-boun...@redfish.com> on behalf of Grant Holland <
> grant.holland...@gmail.com>
> *Sent:* Tuesday, August 8, 2017 4:51:18 PM
> *To:* The Friday Morning Applied Complexity Coffee Group; glen ☣
> *Subject:* Re: [FRIAM] Future of humans and artificial intelligence
>
>
> Thanks for throwing in on this one, Glen. Your thoughts are
> ever-insightful. And ever-entertaining!
>
> For example, I did not know that von Neumann put forth a set theory.
>
> On the other hand... evolution *is* stochastic. (You actually did not
> disagree with me on that. You only said that the reason I was right was
> another one.) A good book on the stochasticity of evolution is "Chance and
> Necessity" by Jacques Monod. (I just finished rereading it for the second
> time. And that proved quite fruitful.)
>
> G.
>
> On 8/8/17 12:44 PM, glen ☣ wrote:
>
>
> I'm not sure how Asimov intended them.  But the three laws is a trope that 
> clearly shows the inadequacy of deontological ethics.  Rules are fine as far 
> as they go.  But they don't go very far.  We can see this even in the 
> foundations of mathematics, the unification of physics, and 
> polyphenism/robustness in biology.  Von Neumann (Burks) said it best when he 
> said: "But in the complicated parts of formal logic it is always one order of 
> magnitude harder to tell what an object can do than to produce the object."  
> Or, if you don't like that, you can see the same perspective in his iterative 
> construction of sets as an alternative to the classical conception.
>
> The point being that reality, traditionally, has shown more expressiveness 
> than any of our rule sets.
>
> There are ways to handle the mismatch in expressivity between reality versus 
> our rule sets.  Stochasticity is the measure of the extent to which a rule 
> set matches a set of patterns.  But Grant's right to qualify that with 
> evolution, not because of the way evolution is stochastic, but because 
> evolution requires a unit to regularly (or sporadically) sync with its 
> environment.
>
> An AI (or a rule-obsessed human) that sprouts fully formed from Zeus' head 
> will *always* fail.  It's guaranteed to fail because syncing with the 
> environment isn't *built in*.  The sync isn't part of the AI's onto- or 
> phylo-geny.
>
>
>
>
> ============================================================
> FRIAM Applied Complexity Group listserv
> Meets Fridays 9a-11:30 at cafe at St. John's College
> to unsubscribe http://redfish.com/mailman/listinfo/friam_redfish.com
> FRIAM-COMIC http://friam-comic.blogspot.com/ by Dr. Strangelove
>
============================================================
FRIAM Applied Complexity Group listserv
Meets Fridays 9a-11:30 at cafe at St. John's College
to unsubscribe http://redfish.com/mailman/listinfo/friam_redfish.com
FRIAM-COMIC http://friam-comic.blogspot.com/ by Dr. Strangelove

Reply via email to