Nick,

Re: your queston about stochastic processes....

Yes, your specific description "AND its last value" is what most uses of "stochastic process" imply. But, technically all that is required to be a "stochastic process" is that each next step in the process is unpredictable, whether or not the outcome of one step influences the outcome of the next. An example of this is the process of flipping a coin several times in a row. Generally, we assume that the outcomes of two adjacent flips are stochastically (or statistically) independent, and that there is no influence between the steps. So, the steps of an independent stochastic process are not dependent on their previous steps.

On the other hand, selecting dinner tonight probably depends on what you had last night, because you would get bored with posole too many nights in a row. And maybe your memory goes back more than just one night, and your selection of dinner tonite is affected by what you had for 2 or more nites before. If your memory goes back only one night, then your "dinner selection process" is a kind of stochastic process called a "Markov process". Markov processes limit their "memory" to just one step. (That keeps the math simpler.)

In any event, stochastic processes whose steps depend on the outcomes of previous steps are "less random" than those that don't, because the earlier steps "give you extra information" that help you narrow down the options and to better predict the future steps - some more than others. So, LEARNING can occur inside of these dependent stochastic processes.

In fact, the mathematics of information theory is all about taking advantage of these dependent (or "conditional") stochastic processes to hopefully predict the outcomes of future steps. The whole thing is based on conditional probability. Info theory uses formulas with names such as joint entropy, conditional entropy, mutual information and entropy rate. These formulas can measure /how much /stochastic dependency is at work in a particular process - i.e how predictable it is. Entropy rate in particular works with conditional stochastic processes and tries to use that "extra information" provided by stochastic dependencies to predict future outcomes.

Re: your "evolution" question... I have been speaking of biological evolution.

HTH

Grant


On 8/9/17 8:47 AM, Nick Thompson wrote:

Hi everybody,

Thanks for your patience as I emerge (hopefully) from post-surgical fog.

I figured I best start my own thread rather than gum up yours.

First. I had always supposed that a stochastic process was one whose value was determined by two factors, a random factor AND it’s last value. So the next step in a random walk is “random” but the current value (it’s present position on a surface, say) is “the result of a stochastic process.” From your responses, and from a short rummage in Wikipedia, I still can’t tell if I am correct or not.

Now remember, you guys, my standard critique of your discourse is that you confuse your models with the facts of nature. What is this “evolution” of which you speak? Unless you tell me otherwise, I will assume you are speaking of the messy biological process of which we are all a result: -- */The alteration of the design of taxa over time/*. Hard to see any way in which that actual process is evidently random. We have to dig deep into the theory that EXPLAINS evolution to find anything that corresponds to the vernacular notion of randomness. There is constraint and predictability all over the place in the evolution I know. Even mutations are predictable. In other words, the randomness of evolution is a creation of your imaginations concerning the phenomenon, not an essential feature of the phenomenon, itself.

So what kind of “evolution” are you guys talking about?

Yes, and forgive me for trolling, a bit. I am trying to wake myself up, here.

nick

Nicholas S. Thompson

Emeritus Professor of Psychology and Biology

Clark University

http://home.earthlink.net/~nickthompson/naturaldesigns/ <http://home.earthlink.net/%7Enickthompson/naturaldesigns/>



============================================================
FRIAM Applied Complexity Group listserv
Meets Fridays 9a-11:30 at cafe at St. John's College
to unsubscribe http://redfish.com/mailman/listinfo/friam_redfish.com
FRIAM-COMIC http://friam-comic.blogspot.com/ by Dr. Strangelove

============================================================
FRIAM Applied Complexity Group listserv
Meets Fridays 9a-11:30 at cafe at St. John's College
to unsubscribe http://redfish.com/mailman/listinfo/friam_redfish.com
FRIAM-COMIC http://friam-comic.blogspot.com/ by Dr. Strangelove

Reply via email to