Statistical Learning from a Regression Perspective considers statistical learning applications when interest centers on the conditional distribution of the response variable, given a set of predictors, and when it is important to characterize how the predictors are related to the response. As a first approximation, this is can be seen as an extension of nonparametric regression. Among the statistical learning procedures examined are bagging, random forests, boosting, and support vector machines. Response variables may be quantitative or categorical. Real applications are emphasized, especially those with practical implications. One important theme is the need to explicitly take into account asymmetric costs in the fitting process. For example, in some situations false positives may be far less costly than false negatives. Another important theme is to not automatically cede modeling decisions to a fitting algorithm. In many settings, subject-matter knowledge should trump formal fitting criteria. Yet another important theme is to appreciate the limitation of one’s data and not apply statistical learning procedures that require more than the data can provide. The material is written for graduate students in the social and life sciences and for researchers who want to apply statistical learning procedures to scientific and policy problems. Intuitive explanations and visual representations are prominent. All of the analyses included are done in R.
“Volpe’s Neurology of the Newborn 7th Edition” has been added to your cart. View cart
Statistical Learning from a Regression Perspective
Author(s): Richard A. Berk
Publisher: Springer
ISBN: 9780387775005
Edition:
$39,99
Delivery: This can be downloaded Immediately after purchasing.
Version: Only PDF Version.
Compatible Devices: Can be read on any device (Kindle, NOOK, Android/IOS devices, Windows, MAC)
Quality: High Quality. No missing contents. Printable
Recommended Software: Check here