Search icon CANCEL
Subscription
0
Cart icon
Cart
Close icon
You have no products in your basket yet
Save more on your purchases!
Savings automatically calculated. No voucher code required
Arrow left icon
All Products
Best Sellers
New Releases
Books
Videos
Audiobooks
Learning Hub
Newsletters
Free Learning
Arrow right icon
Arrow up icon
GO TO TOP
Hands-On Ensemble Learning with R

You're reading from  Hands-On Ensemble Learning with R

Product type Book
Published in Jul 2018
Publisher Packt
ISBN-13 9781788624145
Pages 376 pages
Edition 1st Edition
Languages
Author (1):
Prabhanjan Narayanachar Tattar Prabhanjan Narayanachar Tattar
Profile icon Prabhanjan Narayanachar Tattar
Toc

Table of Contents (17) Chapters close

Hands-On Ensemble Learning with R
Contributors
Preface
1. Introduction to Ensemble Techniques 2. Bootstrapping 3. Bagging 4. Random Forests 5. The Bare Bones Boosting Algorithms 6. Boosting Refinements 7. The General Ensemble Technique 8. Ensemble Diagnostics 9. Ensembling Regression Models 10. Ensembling Survival Models 11. Ensembling Time Series Models 12. What's Next?
Bibliography Index

References


Abraham, B. and Ledolter, J. (https://onlinelibrary.wiley.com/doi/book/10.1002/9780470316610), 1983. Statistical methods for forecasting. J. Wiley

Andersen, P.K., Klein, J.P. and Rosthøj, S., (https://doi.org/10.1093/biomet/90.1.15) 2003. Generalised linear models for correlated pseudo-observations, with applications to multi-state models. Biometrika, 90(1), pp.15-27.

Berk, R.A., (https://www.springer.com/in/book/9783319440477) 2016. Statistical learning from a regression perspective, Second Edition. New York: Springer.

Bou-Hamad, I., Larocque, D. and Ben-Ameur, H., (https://projecteuclid.org/euclid.ssu/1315833185) 2011. A review of survival trees. Statistics Surveys, 5, pp.44-71.

Box, G.E., Jenkins, G.M., Reinsel, G.C. and Ljung, G.M., (https://onlinelibrary.wiley.com/doi/book/10.1002/9781118619193)2015. Time series analysis: forecasting and control. John Wiley & Sons.

Breiman, L., (https://link.springer.com/article/10.1007/BF00058655) 1996. Bagging predictors. Machine learning, 24(2), pp.123-140.

Breiman, L., Friedman, J.H., Olshen, R.A. and Stone, C.J., (https://www.taylorfrancis.com/books/9781351460491) 1984. Classification and regression trees. Routledge.

Broemeling, L.D., (https://www.crcpress.com/Bayesian-Methods-for-Measures-of-Agreement/Broemeling/p/book/9781420083415)2009. Bayesian methods for measures of agreement. Chapman and Hall/CRC.

Bühlmann, P. and Van De Geer, S., (https://www.springer.com/in/book/9783642201912) 2011. Statistics for high-dimensional data: methods, theory and applications. Springer Science & Business Media.

Chatterjee, S. and Hadi, A.S., (https://www.wiley.com/en-us/Regression+Analysis+by+Example%2C+5th+Edition-p-9780470905845) 2012. Regression Analysis by Example, Fifth edition. John Wiley & Sons.

Ciaburro, G., (https://www.packtpub.com/big-data-and-business-intelligence/regression-analysis-r) 2018. Regression Analysis with R, Packt Publishing Ltd.

Cleveland, R.B., Cleveland, W.S., McRae, J.E. and Terpenning, I., (http://www.nniiem.ru/file/news/2016/stl-statistical-model.pdf) 1990. STL: A Seasonal-Trend Decomposition. Journal of Official Statistics, 6(1), pp.3-73.

Cox, D.R., (https://eclass.uoa.gr/modules/document/file.php/MATH394/Papers/%5BCox(1972)%5D Regression Models and Life Tables.pdf) 1972. Regression models and life-tables. Journal of the Royal Statistical Society. Series B (Methodological), 34, pp. 187-220

Cox, D.R., (https://academic.oup.com/biomet/article-abstract/62/2/269/337051) 1975. Partial likelihood. Biometrika, 62(2), pp.269-276.

Dixit, A., (https://www.packtpub.com/big-data-and-business-intelligence/ensemble-machine-learning)2017. Ensemble Machine Learning: A beginner's guide that combines powerful machine learning algorithms to build optimized models. Packt Publishing Ltd.

Draper, N.R. and Smith, H., (https://onlinelibrary.wiley.com/doi/book/10.1002/9781118625590)1999/2014. Applied regression analysis, (Vol. 326). John Wiley & Sons.

Efron, B. (https://projecteuclid.org/download/pdf_1/euclid.aos/1176344552) 1979. Bootstrap methods (https://link.springer.com/chapter/10.1007/978-1-4612-4380-9_41): another look at the jackknife, The Annals of Statistics, 7, 1-26.

Efron, B. and Hastie, T., 2016. (https://web.stanford.edu/~hastie/CASI_files/PDF/casi.pdf) Computer age statistical inference (Vol. 5). Cambridge University Press.

Efron, B. and Tibshirani, R.J., (https://www.crcpress.com/An-Introduction-to-the-Bootstrap/Efron-Tibshirani/p/book/9780412042317) 1994. An introduction to the bootstrap. CRC press.

Friedman, J.H., Hastie, T. and Tibshirani, R. (https://projecteuclid.org/download/pdf_1/euclid.aos/1016218223)2001. Greedy function approximation: A gradient boosting machine. Annals of Statistics, 29(5):1189–1232.

Gordon, L. and Olshen, R.A., (https://europepmc.org/abstract/med/4042086)1985. Tree-structured survival analysis. Cancer treatment reports, 69(10), pp.1065-1069.

Hastie, T., Tibshirani, R., and Friedman, J. (https://www.springer.com/in/book/9780387848570) ,2009., The Elements of Statistical Learning, Second Edition, Springer.

Haykin, S.S, 2009. (https://www.pearson.com/us/higher-education/program/Haykin-Neural-Networks-and-Learning-Machines-3rd-Edition/PGM320370.html) Neural networks and learning machines (Vol. 3). Upper Saddle River, NJ, USA:: Pearson.

Kalbfleisch, J.D. and Prentice, R.L. (https://onlinelibrary.wiley.com/doi/abs/10.2307/3315078), 2002. The statistical analysis of failure time data. John Wiley & Sons.

Kuncheva, L.I., (https://www.wiley.com/en-us/Combining+Pattern+Classifiers%3A+Methods+and+Algorithms%2C+2nd+Edition-p-9781118315231) 2014. Combining pattern classifiers: methods and algorithms. Second Edition. John Wiley & Sons.

LeBlanc, M. and Crowley, J., (https://www.jstor.org/stable/2532300)1992. Relative risk trees for censored survival data. Biometrics, pp.411-425.

Lee, S.S. and Elder, J.F., (http://citeseerx.ist.psu.edu/viewdoc/download;jsessionid=6B151AAB29C69A4D4C35C8C4BBFC67F5?doi=10.1.1.34.1753&rep=rep1&type=pdf) 1997. Bundling heterogeneous classifiers with advisor perceptrons. White Paper.

Mardia, K. , Kent, J., and Bibby, J.M.., (https://www.elsevier.com/books/multivariate-analysis/mardia/978-0-08-057047-1) 1979. Multivariate analysis. Academic Press.

Montgomery, D.C., Peck, E.A. and Vining, G.G., (https://www.wiley.com/en-us/Introduction+to+Linear+Regression+Analysis%2C+5th+Edition-p-9781118627365) 2012. Introduction to linear regression analysis (Vol. 821). John Wiley & Sons.

Perrone, M.P., and Cooper, L.N., (https://www.worldscientific.com/doi/abs/10.1142/9789812795885_0025)1993. When Networks Disagree: Ensemble Methods for Hybrid Neural Networks. In Mammone, R.J. (Ed.), Neural Networks for Speech and Image Processing. Chapman Hall.

Ripley, B.D., (http://admin.cambridge.org/fk/academic/subjects/statistics-probability/computational-statistics-machine-learning-and-information-sc/pattern-recognition-and-neural-networks)2007. Pattern recognition and neural networks. Cambridge university press.

Quenouille, M.H., (https://www.cambridge.org/core/journals/mathematical-proceedings-of-the-cambridge-philosophical-society/article/approximate-tests-of-correlation-in-timeseries-3/F6D24B2A8574F1716E44BE788696F9C7) 1949, July. Approximate tests of correlation in time-series 3. In Mathematical Proceedings of the Cambridge Philosophical Society (Vol. 45, No. 3, pp. 483-484). Cambridge University Press.

Quinlan, J. R. (1993), (https://www.elsevier.com/books/c45/quinlan/978-0-08-050058-4) C4.5: Programs for Machine Learning, Morgan Kaufmann.

Ridgeway, G., Madigan, D. and Richardson, T., (http://dimacs.rutgers.edu/archive/Research/MMS/PAPERS/BNBR.pdf) 1999, January. Boosting methodology for regression problems. In AISTATS.

Schapire, R.E. and Freund, Y., (http://dimacs.rutgers.edu/archive/Research/MMS/PAPERS/BNBR.pdf) 2012. Boosting: Foundations and algorithms. MIT press.

Seni, G. and Elder, J.F., (https://www.morganclaypool.com/doi/abs/10.2200/S00240ED1V01Y200912DMK002)2010. Ensemble methods in data mining: improving accuracy through combining predictions. Synthesis Lectures on Data Mining and Knowledge Discovery, 2(1), pp.1-126.

Tattar, P.N., Ramaiah, S. and Manjunath, B.G., (https://onlinelibrary.wiley.com/doi/book/10.1002/9781119152743) 2016. A Course in Statistics with R. John Wiley & Sons.

Tattar, P.N., 2017. (https://www.packtpub.com/big-data-and-business-intelligence/statistical-application-development-r-and-python-second-edition) Statistical Application Development with R and Python. Packt Publishing Ltd.

Tattar, P., Ojeda, T., Murphy, S.P., Bengfort, B. and Dasgupta, A., (https://www.packtpub.com/big-data-and-business-intelligence/practical-data-science-cookbook-second-edition) 2017. Practical Data Science Cookbook. Packt Publishing Ltd.

Zhang, H. and Singer, B.H., (https://www.springer.com/in/book/9781441968234)2010. Recursive partitioning and applications. Springer Science & Business Media.

Zemel, R.S. and Pitassi, T., (http://papers.nips.cc/paper/1797-a-gradient-based-boosting-algorithm-for-regression-problems.pdf )2001. A gradient-based boosting algorithm for regression problems. In Advances in neural information processing systems (pp. 696-702).

Zhou, Z.H., (https://www.crcpress.com/Ensemble-Methods-Foundations-and-Algorithms/Zhou/p/book/9781439830031)2012. Ensemble methods: foundations and algorithms. Chapman and Hall/CRC.

lock icon The rest of the chapter is locked
Register for a free Packt account to unlock a world of extra content!
A free Packt account unlocks extra newsletters, articles, discounted offers, and much more. Start advancing your knowledge today.
Unlock this book and the full library FREE for 7 days
Get unlimited access to 7000+ expert-authored eBooks and videos courses covering every tech area you can think of
Renews at $15.99/month. Cancel anytime