Statistical Analysis of Intraday Data - Any worthwhile resources?

Discussion in 'Strategy Building' started by Howard, Oct 31, 2015.

  1. dom993

    dom993

    Thank you, and back to the main topic ... did you find anything of interest in that presentation? I am willing to clarify anything discussed in it.
     
    #21     Nov 2, 2015
  2. Howard

    Howard

    Well, like you said, the audio was really bad. I skimmed parts of it, but it didn't seem to be similar to what I'm doing or trying to do. But perhaps I'll give it another go some other day when I have better time.

    It could perhaps seem like I'm a complete newbie starting from scratch, but that's not the situation here. I've spent a huge amount of hours the last year developing my system after a break from the markets. I have extensive amounts of experience with the markets prior to that as well.

    The funny thing is that my system or model seems really simple now that it's (almost) done, but there is a huge amount of work and thinking behind it's logic and operations. :)

    One of the things I've been curious about is for example how much data to look at and how to incorporate data that deviates from what's 'normal'. Like for instance the financial crisis period.

    I'm also mostly looking at patterns based on the last trading day and the day before, but I'm also curious about whether I should add more days, such that I'm looking at for instance 3 days in sequence.

    Anyway, I've realized that these are questions I can answer myself by doing backtesting and research. In fact, I figured out a way to include financial crisis type of data while 'discriminating' it, so to speak.

    I'm not sure if anyone else has experienced this, but I tend to ask questions about where I'm stuck and often the answers just pop up by themselves. Like my brain is working subconsciously while I'm doing other stuff most of the time. Pretty cool. :)
     
    #22     Nov 4, 2015
  3. dom993

    dom993

    One of the challenges for us, trading system developers, is to assess the quality of our work, and especially the likelihood that forward performance will match or come close to backtest performance.

    Most anyone can datamine a good system in backtest, and see it fail to make any money in forward testing. My belief is that the traditional credo re. keeping 1/3 of the data for forward performance assessment is fundamentally flawed (as well as any walkforward optimizations), since any system not performing good enough through these tests is rejected, talk about selection bias here.

    You cannot evaluate your system quality based on its p&l, not even its sharpe ratio, because these are too easily engineered through datamining (aka over-optimization or overfit). The "solution" is called null-hypothesis testing (which is about measuring how far your trade series is from random selection), but you have to find a proper way to do that (which I very much described in that webinar) ... try as hard as you can to understand that aspect because it is way more important than a p/f of 1.5 vs 2+ in your backtest results.

    If you don't mind me posting some backtest results here, I will show you the best pattern I have ever identified - not best according to p&l & other classic metrics, but best as the further away from a random selection. That pattern isn't highly profitable per trade, but it the least susceptible of dramatic counter-performance in forward execution.

    (I'll wait for your consent to post that backtest)
     
    #23     Nov 4, 2015
  4. The Machine by CG3 is all u need to know
     
    #24     Nov 4, 2015
  5. Howard

    Howard

    Knock yourself out. I don't own this thread, so everyone can post whatever they want. : )
     
    #25     Nov 5, 2015
  6. Howard

    Howard

    What's that?
     
    #26     Nov 5, 2015
  7. #27     Nov 5, 2015
  8. dom993

    dom993

    So here are 2 patterns, both trading pullbacks in the direction of the last complete trend.

    Both are "naked" here, meaning only the core-pattern is used, no additional filters.

    Core-pattern vs. Filter :
    • Core-pattern is always an AND of single-metric criterias, each criteria can only be to accept or reject a continuous range of values for that metric.
    • Filters can use more than 1 metric (AND) to reject setups - this amounts to an OR condition to accept setups, which is prone to overfit (hence I always measure Filters quality on the entire market-model space, not on the setup space).
    One of the 2 patterns is 5 stdev from "random", the other one is 6 stdev from "random". There is no doubt that the one at 6 stdev has better statistical significance than the other one, but can you tell which is which from these backtests ?

    Pattern X :

    141-6.no filters.png

    Pattern Y :

    351-v1.no filters.png
     
    #28     Nov 5, 2015
  9. dom993

    dom993

    Sorry I should have indicated that using this market-model, "random" pullback entries yield ~62% win-rate, with avg net/trade ~ -$10 after commissions & slippage
     
    #29     Nov 5, 2015
  10. dom993

    dom993

    and the winner is ...

    ... pattern Y, which is 6 stdev from the market model on its sample size of actually 2250 (even though only 2150 show-up in this backtest), despite its lower win%, lower avg.net / trade, and higher historical max DD, vs pattern X.

    I am going to digress a little and rant about out-of-sample testing (also applicable to walk forward) ... say you have a system with 500 trades in-sample, and you test it out-of-sample which yields 200 trades, and find out the results out of sample are quite different than in-sample (to make things simple, let's say that your primary concern is that the win-rate goes from 62% in-sample to 58% out-of-sample). What does that tell you?

    My answer is, if you just look at those results, it tells you nothing else than the win% is different. There are 2 key aspects that you would need to find answers for, to make any useful conclusion:

    1. How does the performance of random-entries compares out-of-sample to in-sample ?
    2. Is the win% difference between in-sample & out-of sample statistically significant, or is it just noise? (if the random entries performance between the 2 periods are "quite" different, you need to account for that difference as well when working on that question)
     
    #30     Nov 8, 2015
    beginner66 and xandman like this.