Discrepancy-based Algorithms for Best-subset Model Selection PDF Download

Are you looking for read ebook online? Search for your book and save it on your Kindle device, PC, phones or tablets. Download Discrepancy-based Algorithms for Best-subset Model Selection PDF full book. Access full book title Discrepancy-based Algorithms for Best-subset Model Selection by Tao Zhang. Download full books in PDF and EPUB format.

Discrepancy-based Algorithms for Best-subset Model Selection

Discrepancy-based Algorithms for Best-subset Model Selection PDF Author: Tao Zhang
Publisher:
ISBN:
Category : Akaike Information Criterion
Languages : en
Pages : 142

Book Description
The selection of a best-subset regression model from a candidate family is a common problem that arises in many analyses. In best-subset model selection, we consider all possible subsets of regressor variables; thus, numerous candidate models may need to be fit and compared. One of the main challenges of best-subset selection arises from the size of the candidate model family: specifically, the probability of selecting an inappropriate model generally increases as the size of the family increases. For this reason, it is usually difficult to select an optimal model when best-subset selection is attempted based on a moderate to large number of regressor variables. Model selection criteria are often constructed to estimate discrepancy measures used to assess the disparity between each fitted candidate model and the generating model. The Akaike information criterion (AIC) and the corrected AIC (AICc) are designed to estimate the expected Kullback-Leibler (K-L) discrepancy. For best-subset selection, both AIC and AICc are negatively biased, and the use of either criterion will lead to overfitted models. To correct for this bias, we introduce a criterion AICi, which has a penalty term evaluated from Monte Carlo simulation. A multistage model selection procedure AICaps, which utilizes AICi, is proposed for best-subset selection. In the framework of linear regression models, the Gauss discrepancy is another frequently applied measure of proximity between a fitted candidate model and the generating model. Mallows' conceptual predictive statistic (Cp) and the modified Cp (MCp) are designed to estimate the expected Gauss discrepancy. For best-subset selection, Cp and MCp exhibit negative estimation bias. To correct for this bias, we propose a criterion CPSi that again employs a penalty term evaluated from Monte Carlo simulation. We further devise a multistage procedure, CPSaps, which selectively utilizes CPSi. In this thesis, we consider best-subset selection in two different modeling frameworks: linear models and generalized linear models. Extensive simulation studies are compiled to compare the selection behavior of our methods and other traditional model selection criteria. We also apply our methods to a model selection problem in a study of bipolar disorder.

Discrepancy-based Algorithms for Best-subset Model Selection

Discrepancy-based Algorithms for Best-subset Model Selection PDF Author: Tao Zhang
Publisher:
ISBN:
Category : Akaike Information Criterion
Languages : en
Pages : 142

Book Description
The selection of a best-subset regression model from a candidate family is a common problem that arises in many analyses. In best-subset model selection, we consider all possible subsets of regressor variables; thus, numerous candidate models may need to be fit and compared. One of the main challenges of best-subset selection arises from the size of the candidate model family: specifically, the probability of selecting an inappropriate model generally increases as the size of the family increases. For this reason, it is usually difficult to select an optimal model when best-subset selection is attempted based on a moderate to large number of regressor variables. Model selection criteria are often constructed to estimate discrepancy measures used to assess the disparity between each fitted candidate model and the generating model. The Akaike information criterion (AIC) and the corrected AIC (AICc) are designed to estimate the expected Kullback-Leibler (K-L) discrepancy. For best-subset selection, both AIC and AICc are negatively biased, and the use of either criterion will lead to overfitted models. To correct for this bias, we introduce a criterion AICi, which has a penalty term evaluated from Monte Carlo simulation. A multistage model selection procedure AICaps, which utilizes AICi, is proposed for best-subset selection. In the framework of linear regression models, the Gauss discrepancy is another frequently applied measure of proximity between a fitted candidate model and the generating model. Mallows' conceptual predictive statistic (Cp) and the modified Cp (MCp) are designed to estimate the expected Gauss discrepancy. For best-subset selection, Cp and MCp exhibit negative estimation bias. To correct for this bias, we propose a criterion CPSi that again employs a penalty term evaluated from Monte Carlo simulation. We further devise a multistage procedure, CPSaps, which selectively utilizes CPSi. In this thesis, we consider best-subset selection in two different modeling frameworks: linear models and generalized linear models. Extensive simulation studies are compiled to compare the selection behavior of our methods and other traditional model selection criteria. We also apply our methods to a model selection problem in a study of bipolar disorder.

Machine Learning Under a Modern Optimization Lens

Machine Learning Under a Modern Optimization Lens PDF Author: Dimitris Bertsimas
Publisher:
ISBN: 9781733788502
Category : Machine learning
Languages : en
Pages : 589

Book Description


Computational Subset Model Selection Algorithms and Applications

Computational Subset Model Selection Algorithms and Applications PDF Author:
Publisher:
ISBN:
Category :
Languages : en
Pages :

Book Description
This dissertation develops new computationally efficient algorithms for identifying the subset of variables that minimizes any desired information criteria in model selection. In recent years, the statistical literature has placed more and more emphasis on information theoretic model selection criteria. A model selection criterion chooses model that "closely" approximates the true underlying model. Recent years have also seen many exciting developments in the model selection techniques. As demand increases for data mining of massive datasets with many variables, the demand for model selection techniques are becoming much stronger and needed. To this end, we introduce a new Implicit Enumeration (IE) algorithm and a hybridized IE with the Genetic Algorithm (GA) in this dissertation. The proposed Implicit Enumeration algorithm is the first algorithm that explicitly uses an information criterion as the objective function. The algorithm works with a variety of information criteria including some for which the existing branch and bound algorithms developed by Furnival and Wilson (1974) and Gatu and Kontoghiorghies (2003) are not applicable. It also finds the "best" subset model directly without the need of finding the "best" subset of each size as the branch and bound techniques do. The proposed methods are demonstrated in multiple, multivariate, logistic regression and discriminant analysis problems. The implicit enumeration algorithm converged to the optimal solution on real and simulated data sets with up to 80 predictors, thus having 280 = 1,208,925,819,614,630,000,000,000 possible subset models in the model portfolio. To our knowledge, none of the existing exact algorithms have the capability of optimally solving such problems of this size.

Feature Engineering and Selection

Feature Engineering and Selection PDF Author: Max Kuhn
Publisher: CRC Press
ISBN: 1351609467
Category : Business & Economics
Languages : en
Pages : 266

Book Description
The process of developing predictive models includes many stages. Most resources focus on the modeling algorithms but neglect other critical aspects of the modeling process. This book describes techniques for finding the best representations of predictors for modeling and for nding the best subset of predictors for improving model performance. A variety of example data sets are used to illustrate the techniques along with R programs for reproducing the results.

Expert Clouds and Applications

Expert Clouds and Applications PDF Author: I. Jeena Jacob
Publisher: Springer Nature
ISBN: 9811925003
Category : Technology & Engineering
Languages : en
Pages : 785

Book Description
The book features original papers from International Conference on Expert Clouds and Applications (ICOECA 2022), organized by GITAM School of Technology, Bangalore, India, during 3–4 February 2022. It covers new research insights on artificial intelligence, big data, cloud computing, sustainability, knowledge-based expert systems. The book discusses innovative research from all aspects including theoretical, practical, and experimental domains that pertain to the expert systems, sustainable clouds, and artificial intelligence technologies.

Basic Guide for Machine Learning Algorithms and Models

Basic Guide for Machine Learning Algorithms and Models PDF Author: Ms.G.Vanitha
Publisher: SK Research Group of Companies
ISBN: 936492469X
Category : Computers
Languages : en
Pages : 191

Book Description
Ms.G.Vanitha, Associate Professor, Department of Information Technology, Bishop Heber College, Tiruchirappalli, Tamil Nadu, India. Dr.M.Kasthuri, Associate Professor, Department of Computer Science, Bishop Heber College, Tiruchirappalli, Tamil Nadu, India.

Subspace, Latent Structure and Feature Selection

Subspace, Latent Structure and Feature Selection PDF Author: Craig Saunders
Publisher: Springer Science & Business Media
ISBN: 3540341374
Category : Computers
Languages : en
Pages : 218

Book Description
Many of the papers in this proceedings volume were presented at the PASCAL Workshop entitled Subspace, Latent Structure and Feature Selection Techniques: Statistical and Optimization Perspectives which took place in Bohinj, Slovenia during February, 23–25 2005.

Interpretable Machine Learning

Interpretable Machine Learning PDF Author: Christoph Molnar
Publisher: Lulu.com
ISBN: 0244768528
Category : Artificial intelligence
Languages : en
Pages : 320

Book Description
This book is about making machine learning models and their decisions interpretable. After exploring the concepts of interpretability, you will learn about simple, interpretable models such as decision trees, decision rules and linear regression. Later chapters focus on general model-agnostic methods for interpreting black box models like feature importance and accumulated local effects and explaining individual predictions with Shapley values and LIME. All interpretation methods are explained in depth and discussed critically. How do they work under the hood? What are their strengths and weaknesses? How can their outputs be interpreted? This book will enable you to select and correctly apply the interpretation method that is most suitable for your machine learning project.

Machine Learning and Data Science Blueprints for Finance

Machine Learning and Data Science Blueprints for Finance PDF Author: Hariom Tatsat
Publisher: "O'Reilly Media, Inc."
ISBN: 1492073008
Category : Computers
Languages : en
Pages : 432

Book Description
Over the next few decades, machine learning and data science will transform the finance industry. With this practical book, analysts, traders, researchers, and developers will learn how to build machine learning algorithms crucial to the industry. You’ll examine ML concepts and over 20 case studies in supervised, unsupervised, and reinforcement learning, along with natural language processing (NLP). Ideal for professionals working at hedge funds, investment and retail banks, and fintech firms, this book also delves deep into portfolio management, algorithmic trading, derivative pricing, fraud detection, asset price prediction, sentiment analysis, and chatbot development. You’ll explore real-life problems faced by practitioners and learn scientifically sound solutions supported by code and examples. This book covers: Supervised learning regression-based models for trading strategies, derivative pricing, and portfolio management Supervised learning classification-based models for credit default risk prediction, fraud detection, and trading strategies Dimensionality reduction techniques with case studies in portfolio management, trading strategy, and yield curve construction Algorithms and clustering techniques for finding similar objects, with case studies in trading strategies and portfolio management Reinforcement learning models and techniques used for building trading strategies, derivatives hedging, and portfolio management NLP techniques using Python libraries such as NLTK and scikit-learn for transforming text into meaningful representations

Best Practices in Quantitative Methods

Best Practices in Quantitative Methods PDF Author: Jason W. Osborne
Publisher: SAGE
ISBN: 1412940656
Category : Social Science
Languages : en
Pages : 609

Book Description
The contributors to Best Practices in Quantitative Methods envision quantitative methods in the 21st century, identify the best practices, and, where possible, demonstrate the superiority of their recommendations empirically. Editor Jason W. Osborne designed this book with the goal of providing readers with the most effective, evidence-based, modern quantitative methods and quantitative data analysis across the social and behavioral sciences. The text is divided into five main sections covering select best practices in Measurement, Research Design, Basics of Data Analysis, Quantitative Methods, and Advanced Quantitative Methods. Each chapter contains a current and expansive review of the literature, a case for best practices in terms of method, outcomes, inferences, etc., and broad-ranging examples along with any empirical evidence to show why certain techniques are better. Key Features: Describes important implicit knowledge to readers: The chapters in this volume explain the important details of seemingly mundane aspects of quantitative research, making them accessible to readers and demonstrating why it is important to pay attention to these details. Compares and contrasts analytic techniques: The book examines instances where there are multiple options for doing things, and make recommendations as to what is the "best" choice—or choices, as what is best often depends on the circumstances. Offers new procedures to update and explicate traditional techniques: The featured scholars present and explain new options for data analysis, discussing the advantages and disadvantages of the new procedures in depth, describing how to perform them, and demonstrating their use. Intended Audience: Representing the vanguard of research methods for the 21st century, this book is an invaluable resource for graduate students and researchers who want a comprehensive, authoritative resource for practical and sound advice from leading experts in quantitative methods.