The book covers parametric methods or linear models that combat under- or over-fitting using techniques such as Lasso and Ridge. It includes complex regression analysis with time series smoothing, decomposition, and forecasting. It takes a fresh look at non-parametric models for binary classification (logistic regression analysis) and ensemble methods such as decision trees, support vector machines, and naive Bayes. It covers the most popular non-parametric method for time-event data (the Kaplan-Meier estimator). It also covers ways of solving classification problems using artificial neural networks such as restricted Boltzmann machines, multi-layer perceptrons, and deep belief networks. The book discusses unsupervised learning clustering techniques such as the K-means method, agglomerative and Dbscan approaches, and dimension reduction techniques such as Feature Importance, Principal Component Analysis, and Linear Discriminant Analysis. And it introduces driverless artificial intelligence using H2O.
After reading this book, you will be able to develop, test, validate, and optimize statistical machine learning and deep learning models, and engineer, visualize, and interpret sets of data.
What You Will Learn
• Design, develop, train, and validate machine learning and deep learning models
• Find optimal hyper parameters for superior model performance
• Improve model performance using techniques such as dimension reduction and regularization
• Extract meaningful insights for decision making using data visualization
Who This Book Is For
Beginning and intermediate level data scientists and machine learning engineers
The book covers parametric methods or linear models that combat under- or over-fitting using techniques such as Lasso and Ridge. It includes complex regression analysis with time series smoothing, decomposition, and forecasting. It takes a fresh look at non-parametric models for binary classification (logistic regression analysis) and ensemble methods such as decision trees, support vector machines, and naive Bayes. It covers the most popular non-parametric method for time-event data (the Kaplan-Meier estimator). It also covers ways of solving classification problems using artificial neural networks such as restricted Boltzmann machines, multi-layer perceptrons, and deep belief networks. The book discusses unsupervised learning clustering techniques such as the K-means method, agglomerative and Dbscan approaches, and dimension reduction techniques such as Feature Importance, Principal Component Analysis, and Linear Discriminant Analysis. And it introduces driverless artificial intelligence using H2O.
After reading this book, you will be able to develop, test, validate, and optimize statistical machine learning and deep learning models, and engineer, visualize, and interpret sets of data.
What You Will Learn
• Design, develop, train, and validate machine learning and deep learning models
• Find optimal hyper parameters for superior model performance
• Improve model performance using techniques such as dimension reduction and regularization
• Extract meaningful insights for decision making using data visualization
Who This Book Is For
Beginning and intermediate level data scientists and machine learning engineers

Data Science Revealed: With Feature Engineering, Data Visualization, Pipeline Development, and Hyperparameter Tuning
252
Data Science Revealed: With Feature Engineering, Data Visualization, Pipeline Development, and Hyperparameter Tuning
252Paperback(1st ed.)
Product Details
ISBN-13: | 9781484268698 |
---|---|
Publisher: | Apress |
Publication date: | 03/07/2021 |
Edition description: | 1st ed. |
Pages: | 252 |
Product dimensions: | 7.01(w) x 10.00(h) x (d) |