Further reading

Perhaps the most wide-ranging and informative tour of Ensembles and ensemble types is provided by the Kaggle competitor, Triskelion, at http://mlwave.com/kaggle-ensembling-guide/.

For discussion of the Netflix Prize-winning model, Pragmatic Chaos, refer to http://www.stat.osu.edu/~dmsl/GrandPrize2009_BPC_BellKor.pdf. For an explanation by Netflix on how changing business contexts rendered that $1M-model redundant, refer to the Netflix Tech blog at http://techblog.netflix.com/2012/04/netflix-recommendations-beyond-5-stars.html.

For a walkthrough on applying random forest ensembles to commercial contexts, with plenty of space given to all-important diagnostic charts and reasoning, consider Arshavir Blackwell's blog at https://citizennet.com/blog/2012/11/10/random-forests-ensembles-and-performance-metrics/.

For further information on random forests specifically, I find the scikit-learn documentation helpful: http://scikit-learn.org/stable/modules/generated/sklearn.ensemble.RandomForestClassifier.html.

A great introduction to gradient-boosted trees is provided within the XGBoost documentation at http://xgboost.readthedocs.io/en/latest/model.html.

For a write-up of Alexander Guschin's entry to the Otto Product Classification challenge, refer to the No Free Hunch blog: http://blog.kaggle.com/2015/06/09/otto-product-classification-winners-interview-2nd-place-alexander-guschin/.

Alexander Minushkin's Jitter test for overfitting is described at https://www.kaggle.com/miniushkin/introducing-kaggle-scripts/jitter-test-for-overfitting-notebook.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset
3.17.162.247