Do Neural Networks overfit less than Decision Trees?

Discussion in 'Strategy Building' started by damien.dualsigma, Oct 28, 2017.

  1. Alice in Wonderland tree is the only thing that wouldnt overfit
     
    #31     Jan 8, 2018
  2. userque

    userque

    Ok. By patterns, I didn't just mean visual patterns. But I understand.

    But again, I'm not a fan of the NN architecture. It forces formulas on data that is not necessarily rooted in formulas.

    For example:

    input....output
    1 ........ 1
    2 ........ 4
    3 ........ 9

    A formula can be found, and it would fit.

    But if the inputs are essentially random numbers and the outputs also seem to be random:

    2343143 ........ 3452452
    456745674 ........ 6789789
    2342 ....9869
    2343143 ........ 3452452
    456745674 ........ 6789789
    2342 ....9869
    2343143 ........ 3452452
    456745674 ........ 6789789
    2342 ....9869

    There is no underlying formula ... but there is an obvious pattern. A NN will generate a formula even for lottery numbers. A NN will complicate the latter example by generating a network/formula that will 'fit' the data.
     
    Last edited: Jan 8, 2018
    #32     Jan 8, 2018
  3. userque

    userque

    So, what is the author of that paper doing today?

    https://www.google.com/search?q=Gen...er&aqs=chrome..69i57&sourceid=chrome&ie=UTF-8

    "Those who can, do. Those who can't, teach."
     
    #33     Jan 8, 2018
    gon likes this.
  4. gon

    gon

    I see that you mean.

    More than a formula what a NN will do is to weigth the coefficients to adjust the triggering of the activation functions to minimize the loss function of the prediction. Isn't that similar to finding KNNs based upon weighted euclidean distances?

    I mean, the number of neurons within each layer is fixed and the activation fuctions are also preconfigured.

    Regards.
     
    #34     Jan 8, 2018
  5. gon

    gon

    #35     Jan 8, 2018
    userque likes this.
  6. userque

    userque

    Not at all.

    Using the same data:

    2343143 ........ 3452452
    456745674 ........ 6789789
    2342 ....9869
    2343143 ........ 3452452
    456745674 ........ 6789789
    2342 ....9869
    2343143 ........ 3452452
    456745674 ........ 6789789
    2342 ....9869

    Suppose we have a new input of 2343144. The kNN will simply output 3452452. It won't try to figure out a formula to match all of the data, and then plug this new input into it.

    And whatever the correct output turns out to be, it would then be assimilated into the ever growing data set.

    These are simple examples. It gets more complicated and in my custom implementation, proprietary.
     
    #36     Jan 8, 2018
    gon likes this.
  7. ironchef

    ironchef

    But isn't that what us dumb humans are doing predicting short term PA with our "brain power"?
     
    #37     Feb 1, 2018
  8. userque

    userque

    After tutoring HS kids in Math and Science during summer breaks; and after debating/explaining things in forums like this one; and after reading 1,000's of posts in the same forums ... I am absolutely clueless as to wth is going on in some human brains. o_O
     
    #38     Feb 1, 2018
  9. Broadly, neither technique is more likely to overfit than the other. Overfitting (and underfitting, for that matter) may be avoided in any type of empirical model by appropriately establishing model complexity. This is usually tested via error resampling (holdout testing, cross-validation or bootstrapping, for instance). In the case of feedforward neural networks, complexity is typically controlled through early stopping of training or varying the number of hidden nodes. Growth of decision trees is most often constrained via pruning, small sample adjustment of test probabilities or requiring a minimum number of samples for splitting.
     
    #39     Mar 29, 2018
  10. This is true when k = 1, but such models can be extremely noisy: When input values change slightly, the model can swing wildly.
     
    #40     Mar 29, 2018