Bagging predictors

From MaRDI portal
Publication:65108

DOI10.1007/bf00058655zbMath0858.68080WikidataQ29013802 ScholiaQ29013802MaRDI QIDQ65108

Leo Breiman, Leo Breiman

Publication date: August 1996

Published in: Machine Learning (Search for Journal in Brave)




Related Items

Ensemble learning from model based trees with application to differential price sensitivity assessment, A review on instance ranking problems in statistical learning, Optimal Bayesian design for model discrimination via classification, A comparative study of machine learning models for predicting the state of reactive mixing, Comparing boosting and bagging for decision trees of rankings, Breast cancer diagnosis using feature extraction and boosted C5.0 decision tree algorithm with penalty factor, Generalized bagging, Recent advances in statistical methodologies in evaluating program for high-dimensional data, An eager splitting strategy for online decision trees in ensembles, Partitioning around medoids clustering and random forest classification for GIS-informed imputation of fluoride concentration data, Rates of convergence for random forests via generalized U-statistics, Modelling and forecasting based on recursive incomplete pseudoinverse matrices, Explainable models of credit losses, Why estimation alone causes Markowitz portfolio selection to fail and what we might do about it, A random forest algorithm to improve the Lee-Carter mortality forecasting: impact on q-forward, ESG score prediction through random forest algorithm, Dynamic ensemble selection based on hesitant fuzzy multiple criteria decision making, Enhanced prediction of anti-tubercular peptides from sequence information using divergence measure-based intuitionistic fuzzy-rough feature selection, Self-triggered control of probabilistic Boolean control networks: a reinforcement learning approach, Ordinal trees and random forests: score-free recursive partitioning and improved ensembles, Approximate dynamic programming for planning a ride-hailing system using autonomous fleets of electric vehicles, Banzhaf random forests: cooperative game theory based random forests with consistency, Cost-sensitive business failure prediction when misclassification costs are uncertain: a heterogeneous ensemble selection approach, Rejoinder on: ``Hierarchical inference for genome-wide association studies: a view on methodology with software, Cascade interpolation learning with double subspaces and confidence disturbance for imbalanced problems, Multi-output parameter-insensitive kernel twin SVR model, A robust outlier control framework for classification designed with family of homotopy loss function, Micromechanics-based surrogate models for the response of composites: a critical comparison between a classical mesoscale constitutive model, hyper-reduction and neural networks, Models as approximations. II. A model-free theory of parametric regression, Induction of classification rules by Gini-index based rule generation, Propositionalization and embeddings: two sides of the same coin, Double random forest, Robust classification via MOM minimization, A Bayesian perspective of statistical machine learning for big data, Random forest with acceptance-rejection trees, Dynamic recursive tree-based partitioning for malignant melanoma identification in skin lesion dermoscopic images, Modeling of time series using random forests: theoretical developments, A model-free Bayesian classifier, A weighted multiple classifier framework based on random projection, Fast greedy \(\mathcal{C} \)-bound minimization with guarantees, A probabilistic classifier ensemble weighting scheme based on cross-validated accuracy estimates, SIRUS: stable and interpretable RUle set for classification, C443: a methodology to see a forest for the trees, A look at robustness and stability of \(\ell_1\)-versus \(\ell_0\)-regularization: discussion of papers by Bertsimas et al. and Hastie et al., Enhancing techniques for learning decision trees from imbalanced data, From data to stochastic models, Reduced rank regression with matrix projections for high-dimensional multivariate linear regression model, Meta-analysis of voice disorders databases and applied machine learning techniques, Machine learning based classification of normal, slow and fast walking by extracting multimodal features from stride interval time series, Fusion of geochemical and remote-sensing data for lithological mapping using random forest metric learning, Forecasting bankruptcy using biclustering and neural network-based ensembles, A comparative study of the leading machine learning techniques and two new optimization algorithms, Two-level regression method using ensembles of trees with optimal divergence, A deep learning semiparametric regression for adjusting complex confounding structures, Extending models via gradient boosting: an application to Mendelian models, Forecast aggregation via recalibration, Ensemble Gaussian mixture models for probability density estimation, Studying the bandwidth in \(k\)-sample smooth tests, A combination selection algorithm on forecasting, Estimating the functional form of a continuous covariate's effect on survival time, GA-Ensemble: a genetic algorithm for robust ensembles, Canonical forest, A multi-loss super regression learner (MSRL) with application to survival prediction using proteomics, Prediction and classification in nonlinear data analysis: something old, something new, something borrowed, something blue, Comparative study of matrix refinement approaches for ensemble clustering, An instance-based learning recommendation algorithm of imbalance handling methods, Neural network modeling of vector multivariable functions in ill-posed approximation problems, The reliability of classification of terminal nodes in GUIDE decision tree to predict the nonalcoholic fatty liver disease, Determining cutoff point of ensemble trees based on sample size in predicting clinical dose with DNA microarray data, Rotation Forests for regression, Forecasting corporate failure using ensemble of self-organizing neural networks, Machine learning feature analysis illuminates disparity between E3SM climate models and observed climate change, Widening: using parallel resources to improve model quality, Batch mode active learning framework and its application on valuing large variable annuity portfolios, A selective overview of deep learning, A MOM-based ensemble method for robustness, subsampling and hyperparameter tuning, Design-unbiased statistical learning in survey sampling, Ensemble learning based on approximate reducts and bootstrap sampling, SVM-boosting based on Markov resampling: theory and algorithm, Adaboost-based ensemble of polynomial chaos expansion with adaptive sampling, Machine learning for credit scoring: improving logistic regression with non-linear decision-tree effects, Fundamental ratios as predictors of ESG scores: a machine learning approach, Sampling from non-smooth distributions through Langevin diffusion, Bootstrapping multiple linear regression after variable selection, An explicit split point procedure in model-based trees allowing for a quick fitting of GLM trees and GLM forests, Developing automated valuation models for estimating property values: a comparison of global and locally weighted approaches, Deep learning and multivariate time series for cheat detection in video games, Uncertainty quantification for honest regression trees, Optimal selection of sample-size dependent common subsets of covariates for multi-task regression prediction, Bagging-enhanced sampling schedule for functional quadratic regression, Hierarchical resampling for bagging in multistudy prediction with applications to human neurochemical sensing, Computational efficiency of bagging bootstrap bandwidth selection for density estimation with big data, Random clustering forest for extended belief rule-based system, High-dimensional correlation matrix estimation for general continuous data with Bagging technique, Explainable online ensemble of deep neural network pruning for time series forecasting, An evaluation of methods to handle missing data in the context of latent variable interaction analysis: multiple imputation, maximum likelihood, and random forest algorithm, Operational research and artificial intelligence methods in banking, Random forest estimation of conditional distribution functions and conditional quantiles, Robust modeling of unknown dynamical systems via ensemble averaged learning, Asymptotic properties of high-dimensional random forests, Generalization error for Tweedie models: decomposition and error reduction with bagging, Krigings over space and time based on latent low-dimensional structures, The Delaunay triangulation learner and its ensembles, Federated learning in side-channel analysis, Cost-sensitive ensemble learning: a unifying framework, COVID-19 pandemic forecasting using CNN-LSTM: a hybrid approach, An empirical study on classification methods for alarms from a bug-finding static C analyzer, Statistics for phylogenetic trees, An analysis of diversity measures, Training a reciprocal-sigmoid classifier by feature scaling-space, Experimental study for the comparison of classifier combination methods, A cooperative constructive method for neural networks for pattern recognition, On bagging and nonlinear estimation, Improved customer choice predictions using ensemble methods, Adaptive text mining: inferring structure from sequences, Wrappers for feature subset selection, Rationale and applications of survival tree and survival ensemble methods, Performance improvement of classifier fusion for batch samples based on upper integral, Pruning of error correcting output codes by optimization of accuracy-diversity trade off, Two-level quantile regression forests for bias correction in range prediction, Probabilistic combination of classification rules and its application to medical diagnosis, Committee polyhedral separability: complexity and polynomial approximation, A new variable importance measure for random forests with missing data, Ensemble classification of paired data, FSR methods for second-order regression models, Interpretable classifiers using rules and Bayesian analysis: building a better stroke prediction model, A simple method for combining estimates to improve the overall error rates in classification, Ensembles of relational classifiers, Parametric bootstrap approximation to the distribution of EBLUP and related prediction intervals in linear mixed models, A lazy bagging approach to classification, A dynamic overproduce-and-choose strategy for the selection of classifier ensembles, Hidden Markov model-based ensemble methods for offline handwritten text line recognition, Augmenting the bootstrap to analyze high dimensional genomic data, Resampling methods for variable selection in robust regression, Counting and enumerating aggregate classifiers, Effective image retrieval using dominant color descriptor and fuzzy support vector machine, The comparison between classification trees through proximity measures, Aggregating classifiers with mathematical programming, Generalised indirect classifiers, Bundling classifiers by bagging trees, Relevance measures for subset variable selection in regression problems based on \(k\)-additive mutual information, Boosting and instability for regression trees, Maximum a posteriori pruning on decision trees and its application to bootstrap bumping, Responder identification in clinical trials with censored data, Stock and bond return predictability: the discrimination power of model selection criteria, Bootstrap estimated true and false positive rates and ROC curve, Standard errors for bagged and random forest estimators, A similarity measure to assess the stability of classification trees, Using boosting to prune double-bagging ensembles, Survival prediction using gene expression data: a review and comparison, Multiclass classification and gene selection with a stochastic algorithm, Multivariate trees for mixed outcomes, Taxonomy for characterizing ensemble methods in classification tasks: a review and annotated bibliography, Bagging constraint score for feature selection with pairwise constraints, Ore grade prediction using a genetic algorithm and clustering based ensemble neural network model, Disparate data fusion for protein phosphorylation prediction, Iterative Boolean combination of classifiers in the ROC space: an application to anomaly detection with HMMs, Spoofing protection for fingerprint scanner by fusing ridge signal and valley noise, Navigating random forests and related advances in algorithmic modeling, Consistency of random survival forests, Learn\(^{++}\).MF: A random subspace approach for the missing feature problem, Information theoretic combination of pattern classifiers, An asymmetric classifier based on partial least squares, Cost-sensitive boosting for classification of imbalanced data, Functional dissipation microarrays for classification, Decision trees using model ensemble-based nodes, EROS: Ensemble rough subspaces, Fuzzy lattice reasoning (FLR) classifier and its application for ambient ozone estimation, Feature selection via sensitivity analysis of SVM probabilistic outputs, Tree-structured model diagnostics for linear regression, Exact bootstrap \(k\)-nearest neighbor learners, Novel approaches to probabilistic neural networks through bagging and evolutionary estimating of prior probabilities, Diversity of ability and cognitive style for group decision processes, A weight-adjusted voting algorithm for ensembles of classifiers, Experience-consistent modeling: regression and classification problems, Prediction-based regularization using data augmented regression, Iterative bias reduction: a comparative study, Consensus hashing, Data preparation using data quality matrices for classification mining, Imitation learning of car driving skills with decision trees and random forests, Parallelizing AdaBoost by weights dynamics, Classifying densities using functional regression trees: applications in oceanology, Classification by ensembles from random partitions of high-dimensional data, Robust variable selection using least angle regression and elemental set sampling, Trimmed bagging, Classification tree analysis using TARGET, A local boosting algorithm for solving classification problems, Empirical characterization of random forest variable importance measures, On properties of predictors derived with a two-step bootstrap model averaging approach -- a simulation study in the linear regression model, Comparisons of titer estimation methods for multiplexed pneumococcal opsonophagocytic Killing assay, Gaussian processes and limiting linear models, Negative correlation in incremental learning, Methods and algorithms of collective recognition, Hybrid cluster ensemble framework based on the random combination of data transformation operators, Using economic and financial information for stock selection, Bayesian bootstrap prediction, Forecasting cancellation rates for services booking revenue management using data mining, Multi-target support vector regression via correlation regressor chains, Locally linear ensemble for regression, Semi-supervised superpixel classification for medical images segmentation: application to detection of glaucoma disease, On the asymptotics of random forests, A nonlinear aggregation type classifier, Evidential calibration of binary SVM classifiers, Fast decorrelated neural network ensembles with random weights, A dynamic model of classifier competence based on the local fuzzy confusion matrix and the random reference classifier, Using the one-versus-rest strategy with samples balancing to improve pairwise coupling classification, Combining association measures for collocation extraction using clustering of receiver operating characteristic curves, Bagging binary and quantile predictors for time series, Automatic segmentation of blood vessels from retinal fundus images through image processing and data mining techniques, Application of random forest survival models to increase generalizability of decision trees: a case study in acute myocardial infarction, Identification of consistent functional genetic modules, A MapReduce-based distributed SVM ensemble for scalable image classification and annotation, Benchmarking state-of-the-art classification algorithms for credit scoring: an update of research, Ensemble of randomized soft decision trees for robust classification, PBoostGA: pseudo-boosting genetic algorithm for variable ranking and selection, An asymptotically optimal kernel combined classifier, Supervised classification and mathematical optimization, Improved nearest neighbor classifiers by weighting and selection of predictors, Neural network ensembles: Immune-inspired approaches to the diversity of components, On the layered nearest neighbour estimate, the bagged nearest neighbour estimate and the random forest method in regression and classification, Tie-respecting bootstrap methods for estimating distributions of sets and functions of eigenvalues, Knee joint vibration signal analysis with matching pursuit decomposition and dynamic weighted classifier fusion, Sparse ensembles using weighted combination methods based on linear programming, Consolidated trees versus bagging when explanation is required, A statistical view of clustering performance through the theory of \(U\)-processes, Small area estimation of the homeless in Los Angeles: an application of cost-sensitive stochastic gradient boosting, Hierarchical multi-label classification using local neural networks, Pairwise meta-rules for better meta-learning-based algorithm ranking, Preference disaggregation and statistical learning for multicriteria decision support: A review, Survey on speech emotion recognition: features, classification schemes, and databases, On hybrid classification using model assisted posterior estimates, Hellinger distance decision trees are robust and skew-insensitive, A bootstrap-based aggregate classifier for model-based clustering, Algorithmic paradigms for stability-based cluster validity and model selection statistical methods, with applications to microarray data analysis, On robust classification using projection depth, A novel ensemble construction method for multi-view data using random cross-view correlation between within-class examples, Greedy optimization classifiers ensemble based on diversity, Parallel hierarchical sampling: a general-purpose interacting Markov chains Monte Carlo algorithm, Recursive partitioning on incomplete data using surrogate decisions and multiple imputation, Gradient-based boosting for statistical relational learning: the relational dependency network case, Mining axiomatic fuzzy set association rules for classification problems, Derivation of an artificial gene to improve classification accuracy upon gene selection, Using ensemble methods to deal with imbalanced data in predicting protein-protein interactions, Bayesian synthesis: combining subjective analyses, with an application to ozone data, Boosting algorithms: regularization, prediction and model fitting, Hierarchical linear support vector machine, Training regression ensembles by sequential target correction and resampling, Semi-supervised classification based on random subspace dimensionality reduction, Information estimators for weighted observations, On a method for constructing ensembles of regression models, Feature selection for \(k\)-means clustering stability: theoretical analysis and an algorithm, Direct kernel perceptron (DKP): ultra-fast kernel ELM-based classification with non-iterative closed-form weight calculation, Long-term time series prediction using OP-ELM, Weight-selected attribute bagging for credit scoring, Unsupervised weight parameter estimation method for ensemble learning, Dynamic selection of generative-discriminative ensembles for off-line signature verification, Coevolutionary learning of neural network ensemble for complex classification tasks, Estimating the algorithmic variance of randomized ensembles via the bootstrap, Generalized random forests, The dropout learning algorithm, Identifiability of a model for discrete frequency distributions with a multidimensional parameter space, A theoretical foundation of portfolio resampling, Discussion on the paper ``Analysis of spatio-temporal mobile phone data: a case study in the metropolitan area of Milan by P. Secchi, S. Vantini and V. Vitelli, A combined work optimization technology under resource constraints with an application to road repair, Predictive learning via rule ensembles, Correlation and variable importance in random forests, Distributed learning for random vector functional-link networks, baguette, BART: Bayesian additive regression trees, Remembering Leo Breiman, Remembrance of Leo Breiman, Node harvest, Looking inside self-organizing map ensembles with resampling and negative correlation learning, Context-adaptive pre-processing scheme for robust speech recognition in fast-varying noise environment, On the construction of mutual fund portfolios: a multicriteria methodology and an application to the Greek market of equity mutual funds, Ensemble strategies for a medical diagnostic decision support system: A breast cancer diagnosis application, Customer base analysis: partial defection of behaviourally loyal clients in a non-contractual FMCG retail setting, A meta-heuristic approach for improving the accuracy in some classification algorithms, Gene feature interference deconvolution, From cluster ensemble to structure ensemble, Random survival forests, Bayesian Weibull tree models for survival analysis of clinico-genomic data, Tree models for difference and change detection in a complex environment, Stabilizing Variable Selection and Regression, Exploratory subgroup analysis in clinical trials by model selection, A hybrid approach to predicting sports results and an accurate rating system, A support vector machine-based ensemble algorithm for breast cancer diagnosis, The lack of cross-validation can lead to inflated results and spurious conclusions: a re-analysis of the MacArthur violence risk assessment study, Ensembling classification models based on phalanxes of variables with applications in drug discovery, A simple measure of conditional dependence, Complete subset regressions, Out-of-bag estimation of the optimal sample size in bagging, Reducing forgeries in writer-independent off-line signature verification through ensemble of classifiers, BEST : A decision tree algorithm that handles missing values, Optimal prediction pools, Multi-Resolution Functional ANOVA for Large-Scale, Many-Input Computer Experiments, Ensemble classification based on generalized additive models, Discussion of big Bayes stories and BayesBag, Evaluating the impact of a grouping variable on job satisfaction drivers, Accurate tree-based missing data imputation and data fusion within the statistical learning paradigm, Unnamed Item, Improving random forest algorithm by Lasso method, Diverse classifier ensemble creation based on heuristic dataset modification, MODEL OF MODELS: A NEW PERSPECTIVE TO DEAL WITH MODEL UNCERTAINTY, The relative performance of ensemble methods with deep convolutional neural networks for image classification, Experimental comparison of functional and multivariate spectral-based supervised classification methods in hyperspectral image, Predicting risk for adverse health events using random forest, A novel bagging approach for variable ranking and selection via a mixed importance measure, SVM–CART for disease classification, A new procedure for resampled portfolio with shrinkaged covariance matrix, Measuring the Algorithmic Convergence of Randomized Ensembles: The Regression Setting, Data Fusion Methods for Integrating Data-driven Hydrological Models, On stability issues in deriving multivariable regression models, Quantifying uncertainty of subsampling-based ensemble methods under a U-statistic framework, Robust Naive Bayes Combination of Multiple Classifications, Feature Transformation for Corporate Tax Default Prediction: Application of Machine Learning Approaches, Extremizing and Antiextremizing in Bayesian Ensembles of Binary-Event Forecasts, Toward Efficient Ensemble Learning with Structure Constraints: Convergent Algorithms and Applications, Fast FSR Variable Selection with Applications to Clinical Trials, Local Linear Forests, Kriging Riemannian Data via Random Domain Decompositions, A Projection Pursuit Forest Algorithm for Supervised Classification, Performance Assessment of High-dimensional Variable Identification, Toward robust early-warning models: a horse race, ensembles and model uncertainty, POINT AND INTERVAL FORECASTS OF DEATH RATES USING NEURAL NETWORKS, Detection of epistatic effects with logic regression and a classical linear regression model, Estimation and Inference of Heterogeneous Treatment Effects using Random Forests, CatBoost — An Ensemble Machine Learning Model for Prediction and Classification of Student Academic Performance, PREDICTION/ESTIMATION WITH SIMPLE LINEAR MODELS: IS IT REALLY THAT SIMPLE?, Optimal threshold determination based on the mean excess plot, Unnamed Item, Self-adaptation of parameters in a learning classifier system ensemble machine, Statistical Problem Classes and Their Links to Information Theory, Robustify Financial Time Series Forecasting with Bagging, Multinomial-based small area estimation of labour force indicators, Truncation level selection in nonparametric regression using Padé approximation, Improving Reliability Estimation for Individual Numeric Predictions: A Machine Learning Approach, MetaBayes: Bayesian Meta-Interpretative Learning Using Higher-Order Stochastic Refinement, Bagging Stochastic Watershed on Natural Color Image Segmentation, Unnamed Item, Unnamed Item, Unnamed Item, Feature Selection for Heterogeneous Ensembles of Nearest-neighbour Classifiers Using Hybrid Tabu Search, Random Forests for Premise Selection, Renewal type bootstrap for increasing degree \(U\)-process of a Markov chain, A random forest-based approach to combining and ranking seasonality tests, Support vector machines on the D-wave quantum annealer, The Use of Fuzzy t-Conorm Integral for Combining Classifiers, Evolution of high-frequency systematic trading: a performance-driven gradient boosting model, Correlated model fusion, Supervised Learning from Microarray Data, Application of “Aggregated Classifiers” in Survival Time Studies, Bagging Tree Classifiers for Glaucoma Diagnosis, An Ensemble Learning Approach Based on Missing-Valued Tables, O2S2 for the Geodata Deluge, Authorship recognition via fluctuation analysis of network topology and word intermittency, Testing SNPs and sets of SNPs for importance in association studies, Debiased Inference on Treatment Effect in a High-Dimensional Model, Unnamed Item, Bagging Versions of Sliced Inverse Regression, A scale-space approach with wavelets to singularity estimation, Unnamed Item, Unnamed Item, Unnamed Item, Smooth Component Analysis as Ensemble Method for Prediction Improvement, Delta Boosting Machine with Application to General Insurance, Foreign exchange market prediction with multiple classifiers, Modification of Random Forest Based Approach for Streaming Data with Concept Drift, Semi-supervised learning using ensembles of multiple 1D-embedding-based label boosting, ON MATHEMATICAL MODELLING OF SYNTHETIC MEASURES, Model Selection in Estimating Equations, Relating Amino Acid Sequence to Phenotype: Analysis of Peptide-Binding Data, ESTIMATING A PARAMETER WHEN IT IS KNOWN THAT THE PARAMETER EXCEEDS A GIVEN VALUE, A Hybrid Approach of Boosting Against Noisy Data, INFERENCE AFTER MODEL AVERAGING IN LINEAR REGRESSION MODELS, Application of Multi-Input Hamacher-ANFIS Ensemble Model on Stock Price Forecast, Distributed Decision‐Tree Induction in Peer‐to‐Peer Systems, Random survival forests for high‐dimensional data, Composite distance metric integration by leveraging multiple experts' inputs and its application in patient similarity assessment, Using data mining to enable integration of wind resources on the power grid, Random forest and variable importance rankings for correlated survival data, with applications to tooth loss, From designing a single neural network to designing neural network ensembles, Discrete-time survival trees, A nearest-neighbor-based ensemble classifier and its large-sample optimality, Correlation-driven framework based on graph convolutional network for clinical disease classification, Novel Aggregate Deletion/Substitution/Addition Learning Algorithms for Recursive Partitioning, Clustering Categorical Data via Ensembling Dissimilarity Matrices, An Algorithm for Generating Individualized Treatment Decision Trees and Random Forests, A Bayesian Random Split to Build Ensembles of Classification Trees, Rough Ensemble Classifier: A Comparative Study, An analysis of parameter adaptation in reactive tabu search, Ensemble survival trees for identifying subpopulations in personalized medicine, Structural Online Learning, Performance-Based Numerical Solver Selection in the Lighthouse Framework, THEORETICAL FOUNDATIONS AND EXPERIMENTAL RESULTS FOR A HIERARCHICAL CLASSIFIER WITH OVERLAPPING CLUSTERS, A combined strategy for multivariate density estimation, Hybrid Classification of High-Dimensional Biomedical Tumour Datasets, APPLYING ECONOMIC MEASURES TO LAPSE RISK MANAGEMENT WITH MACHINE LEARNING APPROACHES, Probability estimation with machine learning methods for dichotomous and multicategory outcome: Theory, Constraint relaxation, cost-sensitive learning and bagging for imbalanced classification problems with outliers, Fusing vantage point trees and linear discriminants for fast feature classification, A new variable selection approach using random forests, Stable prediction in high-dimensional linear models, Ranking in evolving complex networks, Bagging strategies for learning planning policies, FuzzyBagging: a novel ensemble of classifiers, Classification by evolutionary ensembles, Consensus analysis of multiple classifiers using non-repetitive variables: diagnostic application to microarray gene expression data, A random forest guided tour, Bayesian multinomial latent variable modeling for fraud and abuse detection in health insurance, LiNearN: a new approach to nearest neighbour density estimator, VR-BFDT: a variance reduction based binary fuzzy decision tree induction method for protein function prediction, Impact of geometric uncertainty on hemodynamic simulations using machine learning, Prediction of FMN-binding residues with three-dimensional probability distributions of interacting atoms on protein surfaces, Projective clustering ensembles, An intelligent-agent-based fuzzy group decision making model for financial multicriteria decision support: The case of credit scoring, Cancer classification using ensemble of neural networks with multiple significant gene subsets, Active learning for logistic regression: an evaluation, Neural network ensemble strategies for financial decision applications, The combination of multiple classifiers using an evidential reasoning approach, Ensemble learning HMM for motion recognition and retrieval by Isomap dimension reduction, Mixing partially linear regression models, Automatic color constancy algorithm selection and combination, A survey of deep network techniques all classifiers can adopt, Ensembles of cost-diverse Bayesian neural learners for imbalanced binary classification, Rough cognitive ensembles, A random forest based approach for predicting spreads in the primary catastrophe bond market, Validating game-theoretic models of terrorism: insights from machine learning, Absolute convergence and error thresholds in non-active adaptive sampling, PAC-Bayesian risk bounds for group-analysis sparse regression by exponential weighting, An embedded model estimator for non-stationary random functions using multiple secondary variables, Hierarchical mixing linear support vector machines for nonlinear classification, Efficient use of data for LSTM mortality forecasting, Arbitrage of forecasting experts, Random forests in count data modelling: an analysis of the influence of data features and overdispersion on regression performance, First order random forests: Learning relational classifiers with complex aggregates, Gleaner: Creating ensembles of first-order clauses to improve recall-precision curves, Training and assessing classification rules with imbalanced data, Optimal experimental design and some related control problems, Risk reduction for nonlinear prediction and its application to the surrogate data test, Combining the requirement information for software defect estimation in design time, Assessing the stability of classification trees using Florida birth data, Computer-intensive rate estimation, diverging statistics and scanning, Neural network ensembles: evaluation of aggregation algorithms, An empirical study of using Rotation Forest to improve regressors, Cross-validated bagged learning, The random electrode selection ensemble for EEG signal classification, Genetic algorithm-based feature set partitioning for classification problems, Building ensemble classifiers using belief functions and OWA operators, Maximum patterns in datasets, An efficient modified boosting method for solving classification problems, Model combination for credit risk assessment: a stacked generalization approach, Selected tree classifier combination based on both accuracy and error diversity, Algorithms for manipulation of level sets of nonparametric density estimates, Classifying G-protein coupled receptors with bagging classification tree, Additive regularization trade-off: fusion of training and validation levels in kernel methods, Network traffic classification based on ensemble learning and co-training, Automatic knowledge learning and case adaptation with a hybrid committee approach, Management of uncertainty in statistical reasoning: the case of regression analysis, Manufacturing lead time estimation using data mining, Optimization of a training set for more robust face detection, Bagging null space locality preserving discriminant classifiers for face recognition, Comments on ``Data science, big data and statistics, Diversification for better classification trees, Estimating residual variance in random forest regression, A novel multi-view learning developed from single-view patterns, A probabilistic model of classifier competence for dynamic ensemble selection, Bagging of density estimators, Approximating high-dimensional infinite-order \(U\)-statistics: statistical and computational guarantees, A long short-term memory ensemble approach for improving the outcome prediction in intensive care unit, Wavelet neural networks: a practical guide, Forecasting financial and macroeconomic variables using data reduction methods: new empirical evidence, A boosting method with asymmetric mislabeling probabilities which depend on covariates, Learning a priori constrained weighted majority votes, Regression conformal prediction with random forests, On the fusion of threshold classifiers for categorization and dimensionality reduction, Classification of repeated measurements data using tree-based ensemble methods, Predictive models for bariatric surgery risks with imbalanced medical datasets, Selective ensemble of SVDDs with Renyi entropy based diversity measure, Rejoinder on: Augmenting the bootstrap to analyze high dimensional genomic data, The method of ensembles of decision trees for the analysis of the electrical activity produced by the human brain, A hierarchy of sum-product networks using robustness, Supervised learning via smoothed Polya trees, Multi-scale rois selection for classifying multi-spectral images, On the interpretation of ensemble classifiers in terms of Bayes classifiers, Estimating a sharp convergence bound for randomized ensembles, Sparse kernel deep stacking networks, On PAC-Bayesian bounds for random forests, Propensity score estimation using classification and regression trees in the presence of missing covariate data, Robust clustering tools based on optimal transportation, Application of the bootstrap method for change points analysis in generalized linear models, Massive datasets and machine learning for computational biomedicine: trends and challenges, Hybrid extreme rotation forest, Regression trees and forests for non-homogeneous Poisson processes, Complexities of convex combinations and bounding the generalization error in classification, Multilabel classification through random graph ensembles, Random projections as regularizers: learning a linear discriminant from fewer observations than dimensions, Local discriminative distance metrics ensemble learning, Structural-break models under mis-specification: implications for forecasting, Domain estimation under informative linkage, Comments on: ``A random forest guided tour, Comments on: ``A random forest guided tour, COMPLETE SUBSET AVERAGING FOR QUANTILE REGRESSIONS, The Impact of Churn on Client Value in Health Insurance, Evaluation Using a Random Forest Under Various Censoring Mechanisms, Partial Least Squares for Heterogeneous Data, Robust ranking by ensembling of diverse models and assessment metrics, Unnamed Item, Ensemble of fast learning stochastic gradient boosting, A model-averaging approach for smoothing spline regression, fried, Joint leaf-refinement and ensemble pruning through \(L_1\) regularization, Quantum amplitude amplification algorithm simulation for prediction of a binary classification problem, ROC‐guided survival trees and ensembles, Machine Learning Surrogate Modeling for Meshless Methods: Leveraging Universal Approximation, Forecasting vector autoregressions with mixed roots in the vicinity of unity, Nonparametric inference for interval data using kernel methods, Regression with missing data, a comparison study of techniques based on random forests, Ensemble Economic Scenario Generators: Unity Makes Strength, A hybrid ensemble method with negative correlation learning for regression, Hierarchical fuzzy regression tree: a new gradient boosting approach to design a TSK fuzzy model, A power-controlled reliability assessment for multi-class probabilistic classifiers, Generalized decision directed acyclic graphs for classification tasks, Supervised Machine Learning Techniques: An Overview with Applications to Banking, Risk aversion in share auctions: Estimating import rents from TRQs in Switzerland, Ensemble Subset Regression (ENSURE): Efficient High-dimensional Prediction, A two‐stage forecasting approach for short‐term intermodal freight prediction, Detecting cell assemblies by NMF-based clustering from calcium imaging data, A framework for machine learning of model error in dynamical systems, An Outer-Product-of-Gradient Approach to Dimension Reduction and its Application to Classification in High Dimensional Space, Discrete IV \(d_G\)-Choquet integrals with respect to admissible orders, Global sensitivity analysis of randomized trials with nonmonotone missing binary outcomes: Application to studies of substance use disorders, A Novel Bayesian Functional Spatial Partitioning Method with Application to Prostate Cancer Lesion Detection Using MRI, Selective Prediction-Set Models with Coverage Rate Guarantees, NNBits: bit profiling with a deep learning ensemble based distinguisher, Inverse attribute‐based optimization with an application in assortment optimization, A hybrid generalized propensity score approach for observational studies, Extrapolation‐based Bandwidth Selectors: A Review and Comparative Study with Discussion on Bivariate Applications, Optimization over decision trees: a case study for the design of stable direct-current electricity networks, BAYESIAN DYNAMIC VARIABLE SELECTION IN HIGH DIMENSIONS, Bayesian regularization for flexible baseline hazard functions in Cox survival models, Breaking the curse of dimensionality in quadratic discriminant analysis models with a novel variant of a Bayes classifier enhances automated taxa identification of freshwater macroinvertebrates, Machine learning in bank merger prediction: a text-based approach, Heuristic methods for stock selection and allocation in an index tracking problem, Predictive Subdata Selection for Computer Models, High-dimensional Canonical Forest, Hierarchically structured task-agnostic continual learning, Random forest pruning techniques: a recent review, A systematic literature review on the use of machine learning in code clone research, Regression trees and ensemble for multivariate outcomes, Comparison of various machine learning algorithms for estimating generalized propensity score, Is there a role for statistics in artificial intelligence?, Supervised classification of curves via a combined use of functional data analysis and tree-based methods, A Survey of Evolution in Predictive Models and Impacting Factors in Customer Churn, A reflected feature space for CART, Weighted bagging: a modification of AdaBoost from the perspective of importance sampling, Bias-corrected random forests in regression, An empirical bias–variance analysis of DECORATE ensemble method at different training sample sizes, Encoding Through Patterns: Regression Tree–Based Neuronal Population Models, RandGA: injecting randomness into parallel genetic algorithm for variable selection, Climate, agriculture, and hunger: statistical prediction of undernourishment using nonlinear regression and data-mining techniques, Comparative analysis of classification methods in determining non-active student characteristics in Indonesia Open University, AN EFFECTIVE BIAS-CORRECTED BAGGING METHOD FOR THE VALUATION OF LARGE VARIABLE ANNUITY PORTFOLIOS, Random forests for time-dependent processes, Optimization of Tree Ensembles, On the independence requirement in Dempster-Shafer theory for combining classifiers providing statistical evidence, Extremely randomized trees, Variable Selection for Marginal Longitudinal Generalized Linear Models, Boosting Insights in Insurance Tariff Plans with Tree-Based Machine Learning Methods, Extremely randomized trees, On the independence requirement in Dempster-Shafer theory for combining classifiers providing statistical evidence, When Errors Become the Rule, Discrete-time survival trees and forests with time-varying covariates, Stochastic gradient boosting., Looking for lumps: boosting and bagging for density estimation., Effect of pruning and early stopping on performance of a boosting ensemble., Improving nonparametric regression methods by bagging and boosting., Generalized additive multi-mixture model for data mining., A statistical approach to growing a reliable honest tree., Statistical significance of the Netflix challenge, Seeing the Forest Through the Trees, Rule Validation of a Meta-classifier Through a Galois (Concept) Lattice and Complementary Means, Recent developments in bootstrap methodology, Bootstrapping phylogenetic trees: theory and methods, Randomized maximum-contrast selection: subagging for large-scale regression, Aggregating classifiers via Rademacher–Walsh polynomials, Censoring Unbiased Regression Trees and Ensembles, Confidence intervals centred on bootstrap smoothed estimators, Unnamed Item, Unnamed Item, Variable selection in Bayesian generalized linear‐mixed models: An illustration using candidate gene case‐control association studies, Classifier Chains: A Review and Perspectives, Appropriate machine learning techniques for credit scoring and bankruptcy prediction in banking and finance: A comparative study, Variable Selection Methods in High-dimensional Regression—A Simulation Study, Reduced forms and weak instrumentation, Bootstrap aggregated classification for sparse functional data, Comments on ``Data science, big data and statistics, Cost-Complexity Pruning of Random Forests, High-Dimensional Data Classification, How Much Telematics Information Do Insurers Need for Claim Classification?, Estimator selection and combination in scalar-on-function regression, On principal components regression, random projections, and column subsampling, An ensemble approach for in silico prediction of Ames mutagenicity, Random average shifted histograms, Accurate ensemble pruning with PL-bagging, Ensemble of a subset of \(k\)NN classifiers, Stable feature selection for biomarker discovery, Ensembles for multi-target regression with random output selections, An approximate likelihood perspective on ABC methods, Distant diversity in dynamic class prediction, Evaluating the importance of different communication types in romantic tie prediction on social media, Subsampling based inference for \(U\) statistics under thick tails using self-normalization, Heuristics of instability and stabilization in model selection, Gradient boosting for high-dimensional prediction of rare events, Should we impute or should we weight? Examining the performance of two CART-based techniques for addressing missing data in small sample research with nonnormal variables, Deriving optimal data-analytic regimes from benchmarking studies, Using the Bayesian Shtarkov solution for predictions, A triplot for multiclass classification visualisation, Noise peeling methods to improve boosting algorithms, Addressing overfitting and underfitting in Gaussian model-based clustering, Grouped variable importance with random forests and application to multiple functional data analysis, A robust AdaBoost.RT based ensemble extreme learning machine, A new ensemble method with feature space partitioning for high-dimensional data classification, Random forest with adaptive local template for pedestrian detection, Optimizing area under the ROC curve using semi-supervised learning, Dynamic classifier aggregation using interaction-sensitive fuzzy measures, Weighted classifier ensemble based on quadratic form, Accelerating difficulty estimation for conformal regression forests, Multi-target regression via input space expansion: treating targets as inputs, Context-based unsupervised ensemble learning and feature ranking, Bootstrapping some GLM and survival regression variable selection estimators, Naive automated machine learning, Algorithm selection on a meta level, Random Forests for Spatially Dependent Data, Fault classification via energy based features of two-dimensional image data, A Neural Approach to Improve the Lee-Carter Mortality Density Forecasts, Efficient permutation testing of variable importance measures by the example of random forests, Use of random forest for assessing the effect of water quality parameters on the biological status of surface waters, Active learning by query by committee with robust divergences, Subsampling based variable selection for generalized linear models, Model averaging for support vector classifier by cross-validation, Bernstein-von Mises theorem and misspecified models: a review, Space-dependent turbulence model aggregation using machine learning, An ensemble EM algorithm for Bayesian variable selection, Reproducible model selection using bagged posteriors, Two efficient selection methods for high‐dimensional <scp>CD‐CAT</scp> utilizing max‐marginals factor from <scp>MAP</scp> query and ensemble learning approach, A new robust classifier on noise domains: bagging of Credal C4.5 trees, Geometry of EM and related iterative algorithms, Mixture of inhomogeneous matrix models for species‐rich ecosystems, Attribute bagging: Improving accuracy of classifier ensembles by using random feature subsets., Double-bagging: Combining classifiers by bootstrap aggregation, Modeling threshold interaction effects through the logistic classification trunk, Analysis of web visit histories. II: Predicting navigation by nested STUMP regression trees, Exploratory machine learning with unknown unknowns, Optimal Nonparametric Inference with Two-Scale Distributional Nearest Neighbors, Large Scale Prediction with Decision Trees, Modelplasticity and abductive decision making, Evaluation of physics constrained data-driven methods for turbulence model uncertainty quantification, A comparison of two dissimilarity functions for mixed-type predictor variables in the \(\delta\)-machine, Dynamic risk prediction triggered by intermediate events using survival tree ensembles, Estimation and inference of treatment effects with \(L_2\)-boosting in high-dimensional settings, Model averaging prediction by \(K\)-fold cross-validation, 2-step gradient boosting approach to selectivity bias correction in tax audit: an application to the VAT gap in Italy, Derandomizing Knockoffs, Linear and nonlinear model predictive control (MPC) for regulating pedestrian flows with discrete speed instructions, Predicting the spatial distribution of stable isotopes in precipitation using a machine learning approach: a comparative assessment of random forest variants, Neighborhood-based cross fitting approach to treatment effects with high-dimensional data, Data-driven decision model based on local two-stage weighted ensemble learning, Meta Clustering for Collaborative Learning, Ensemble learning for the partial label ranking problem, Robust regression using biased objectives, Solute transport prediction in heterogeneous porous media using random walks and machine learning, Back-propagating errors through artificial neural networks for variable selection, Bootstrap bias corrections for ensemble methods, Bayesian additive regression trees using Bayesian model averaging, The online performance estimation framework: heterogeneous ensemble learning for data streams, Wide consensus aggregation in the Wasserstein space. Application to location-scatter families, Boosting imbalanced data learning with Wiener process oversampling, Comparison of machine learning methods for copper ore grade estimation, Unnamed Item, Unnamed Item, Unnamed Item, An incremental classification algorithm for mining data with feature space heterogeneity, Bagged Filters for Partially Observed Interacting Systems, Adaptive linear and normalized combination of radial basis function networks for function approximation and regression, Vote counting measures for ensemble classifiers., Constructing support vector machine ensemble., Bayesian model averaging: A tutorial. (with comments and a rejoinder)., A concrete statistical realization of Kleinberg's stochastic dicrimination for pattern recognition. I: Two-class classification, Shape constraints in economics and operations research, Logical analysis of data as a tool for the analysis of probabilistic discrete choice behavior, Fitting random cash management models to data, Noisy replication in skewed binary classification., Multicriteria classification and sorting methods: A literature review, A hierarchical classification strategy for digital documents, Efficient data reconciliation, symbolicDA, A geometric approach to leveraging weak learners, Ensembling neural networks: Many could be better than all, Interpreting neural-network results: a simulation study., Bounding the generalization error of convex combinations of classifiers: Balancing the dimensionality and the margins., Generalization error of combined classifiers., Least angle regression. (With discussion), Generalization bounds for averaged classifiers, Is it worth generating rules from neural network ensembles?, Population theory for boosting ensembles., On the Bayes-risk consistency of regularized boosting methods., Optimal aggregation of classifiers in statistical learning., Learning hybrid neuro-fuzzy classifier models from data: to combine or not to combine?, Stability and scalability in decision trees, Sensory analysis in the food industry as a tool for marketing decisions, Predicting partial customer churn using Markov for discrimination for modeling first purchase sequences, Boosting random subspace method, Exploiting unlabeled data to enhance ensemble diversity, A review of survival trees, Applying randomness effectively based on random forests for classification task of datasets of insufficient information, Statistical uncertainty estimation using random forests and its application to drought forecast, Multiple-view multiple-learner active learning, Regularization in skewed binary classification, Credit spread approximation and improvement using random forest regression, Algorithms for drug sensitivity prediction, Forecasting using random subspace methods, Generalized random shapelet forests, Cluster ensembles: a survey of approaches with recent extensions and applications, A nearest neighbour extension to project duration forecasting with artificial intelligence, Exploring the sources of uncertainty: why does bagging for time series forecasting work?, Improving bagging performance through multi-algorithm ensembles, Classification of multivariate time series and structured data using constructive induction, Cost-sensitive learning and decision making revisited, Correcting classifiers for sample selection bias in two-phase case-control studies, High-dimensional inference: confidence intervals, \(p\)-values and R-software \texttt{hdi}, A novel margin-based measure for directed hill climbing ensemble pruning, Stable graphical model estimation with random forests for discrete, continuous, and mixed variables, Arcing classifiers. (With discussion), Boosting the margin: a new explanation for the effectiveness of voting methods, Additive logistic regression: a statistical view of boosting. (With discussion and a rejoinder by the authors), On weak base hypotheses and their implications for boosting regression and classification, Analyzing bagging, Discriminant feature extraction using empirical probability density estimation and a local basis library, Partitioning algorithms and combined model integration for data mining, Combining statistical and reinforcement learning in rule-based classification