Stress testing is the activity of subjecting a potential investment strategy to plausible volatility from the market. There is a record of what the market has done in the past, so one way to do stress testing is to do retrodiction on many different past periods to see how well the strategy performs. One doesn't cheat in that testing (peek into the known future to train the model), beyond the recognition that the strategy must integrate knowledge of how markets work. That kind of circularity exists in science. Where does a hypothesis come from if not from past experience?
A limitation of this kind of protocol is that every bit of data one uses for training must be kept separate from the data one uses for testing. While cross validation is possible -- using different training and testing subsets -- at the end of the day history only tells you what has happened not what could happen. Generative machine learning techniques can learn many-body joint probabilities of market events like price moves of different instruments. Using these techniques one can generate not just a tape recording of what happened, but plausible looking trading patterns that share similar patterns -- like might occur in the future. For ABM to be better, it seems to me, there has to be the possibility that a modeler can imagine subtleties of individual behaviors that, when coupled together, create phenomena that has never been seen on markets where equities can trade dozens of times a second. Such a simulation could involve hundreds, or even thousands of types of traders each with their own evolving motivations. Never mind the difficulty of implementing such a model, how could those rules be learned if not inferred from data? How does one presume to know the motivations and resources of traders that often take active measures to hide themselves, e.g. disguise their price impact? Perhaps this is more of an explanatory (academic) effort, positing that stress testing is possible using relatively simple models. That's a nice hypothesis, maybe it even could explain a lot of variation, and the result of that would be satisfying, and a contribution to theory. I expect practitioners have a different, higher, bar. Marcus On 11/15/19, 10:30 AM, "Friam on behalf of uǝlƃ ☣" <friam-boun...@redfish.com on behalf of geprope...@gmail.com> wrote: Since Marcus hasn't answered, I think it's important that he pre-pended ML with "generative". There's plenty to argue about with respect to that word (e.g. my complaint that EricC _thinned_ relativism). But in my own work, I've made the somewhat hand-waving argument that mathematical models (e.g. systems of ODEs) are thin and component-based models are thick. But, I'll take the opportunity, here, to argue against myself (and against you, Steve 8^) that not only can mathematical models be "a little bit thick", but that models induced into combined-but-separable probability distributions are also "a little bit thick". For math models, each term of some equation kinda-sorta represents a component of the model and then various operators are used to integrate those components (+ and - are the most boring). At the next layer, different constraints, mechanisms, components might be modeled by entirely different equations that have to be solved as a system. So, even in this brief conception, it seems clear that these models *can be* mechanistic ... i.e. explanatory, at least to some extent. The same might be said of an induction method that produces, say, bifurcated components. If you find 2 modes in some output, then it seems reasonable that there might be 2 mechanisms at work. So, in direct response to your response. Had you said the point of agent models is to yield *more* explanatory results than a generative ML classifier, I don't think there's much room to argue. We could turn the tables and argue that agent models might be more explanatory, but they'll be less predictive. So, maybe the total power is similar and we should all use *both*. (That's what I argue to my clients ... but it's rarely done because the skill sets are a bit different and it's more expensive. [sigh]) On 11/14/19 5:56 PM, Steven A Smith wrote: > > On 11/14/19 6:10 PM, Marcus Daniels wrote: >> >> Generative machine learning seems a heck of a lot easier than ABMs for stress testing. >> > Agent-based models, used in fields from biology to sociology, are bottom-up, simulating the messy interactions of hundreds and even millions of agents—human cells or attitudes or financial firms—to explain the behavior of a complex system. > > I think the point of Agent Models is to yield *explanatory* not just *predictive results? -- ☣ uǝlƃ ============================================================ FRIAM Applied Complexity Group listserv Meets Fridays 9a-11:30 at cafe at St. John's College to unsubscribe http://redfish.com/mailman/listinfo/friam_redfish.com archives back to 2003: http://friam.471366.n2.nabble.com/ FRIAM-COMIC http://friam-comic.blogspot.com/ by Dr. Strangelove ============================================================ FRIAM Applied Complexity Group listserv Meets Fridays 9a-11:30 at cafe at St. John's College to unsubscribe http://redfish.com/mailman/listinfo/friam_redfish.com archives back to 2003: http://friam.471366.n2.nabble.com/ FRIAM-COMIC http://friam-comic.blogspot.com/ by Dr. Strangelove