Questions tagged [boosting]

Boosting is a machine learning ensemble meta-algorithm in supervised learning, and a family of machine learning algorithms that convert weak learners to strong ones. Also: Boosting is the process of enhancing the relevancy of a document or field

From [the docs]:

"Boosting" is a machine learning ensemble meta-algorithm for primarily reducing bias, and also variance in supervised learning, and a family of machine learning algorithms that convert weak learners to strong ones.

Also:

From the docs:

Boosting is the process of enhancing the relevancy of a document or field. Field level mapping allows to define an explicit boost level on a specific field. The boost field mapping (applied on the root object) allows to define a boost field mapping where its content will control the boost level of the document.

181 questions
0
votes
1 answer

Boosting Classification Tree gbm character variable

I am trying to boost my classification tree and was trying to use the gbm package. My variables are character variables of 'N' and 'Y', so the gbm.fit brings an error that my variable "is not of type numeric, ordered, or factor." Is there a way to…
Joseph Kim
  • 13
  • 2
  • 4
0
votes
0 answers

Segmentation fault when defining a custom loss in lightgbm

I've been trying the following custom loss for a regression task in Lightgbm and boosting type = "goss", and I'm getting a segmentation fault: def hm(y_true, y_pred): residual = (y_true - y_pred).astype("float") residual_abs =…
0
votes
0 answers

Am I using xgboost() correctly (in R)?

I'm a beginner with machine learning (and also R). I've figured out how to run some basic linear regression, elastic net, and random forest models in R and have gotten some decent results for a regression project (with a continuous dependent…
user10605319
0
votes
2 answers

Concept Behind the Ensemble Learning?

I have tried to learn the ensemble ML algorithms like Bagging and Boosting. But I not clear about that why in case of Bagging algorithm we have to consider Decision trees parallelly as base model and in case of Boosting have to take Decision tree…
0
votes
1 answer

Random Forest Cp is 0

I am trying to run a randomforest model for price prediction, and when I run tuneGrid in randomforest, I found that the bestTune cp is 0.00, is this possible?
0
votes
1 answer

Sklearn AdaBooster and Base estimators

I want to use the sklearn AdaBoostRegressor with different base estimators. The general AdaBoost introduction does not help too much, since they use the DecisionTreeClassifier Where do I find a list of all base possible base estimators? Could I…
Uwe.Schneider
  • 1,112
  • 1
  • 15
  • 28
0
votes
1 answer

CatBoost RandomizedSearch stopping and finding out hyperparameters for it

So I started running CatBoost's RandomizedSearch implementation and so far as I see from the log (which I know from verbose=1, down below is the log), there is one model that is the best one for a really long time, and I cannot wait anymore for it…
GreatDuke
  • 214
  • 1
  • 10
0
votes
3 answers

In what scenario bagging can be used over boosting?

I am new to data science and so far i have learnt that bagging only reduces high variance but boosting reduces both variance and bias and thus increasing the accuracy for both train and test cases. I understand the functioning of both. Seems like…
0
votes
1 answer

Reversing machine learning models to get particular features

I am trying to model a process. My input data includes certain features and measurements about the product. I built Random forests and Gradient boosting models in Python, and got good results. I am now trying to determine which features and…
Derek Langley
  • 172
  • 1
  • 11
0
votes
4 answers

low training (~64%) and test accuracy (~14%) with 5 different models

Im struggling to find a learning algorithm that works for my dataset. I am working with a typical regressor problem. There are 6 features in the dataset that I am concerned with. There are about 800 data points in my dataset. The features and the…
0
votes
1 answer

Caret: family specification in glmboost doesn't work

I'm trying to run a boosted robust regression on Caret (with the Huber family), however I get an error when training the model: library(caret) X <- rnorm(300, 0, 100) Y <- rnorm(300, 0, 100000) data <- cbind(X,Y) model <- train(Y~X,…
0
votes
1 answer

Solr 7 boosting

Using SOLR 7.X I am looking at boosting a search based on a SKU match select?fl=SKU&q=text:234^1 OR SKU:A234-TRIM-WH^10 with result: 0 0
0
votes
0 answers

Visualize components of data-generating process in R

I try to replicate this figure with the true underlying function given also there (see also code below). I was wondering how the author came up with this (at first glance easy to replicate) figure. If I look e.g. at the first component of (11)…
0
votes
1 answer

LogitBoost requires the base estimator to be a regressor

I have a dataset that all the values for each feature are numeric, even the class/label column. In boosting algorithms implemented in python (like logitboost, adaboost, gradientboosting), other than the preset base estimators (or weak learners,…
user10726240
0
votes
1 answer

How to simulate bias and variance of a smoothing spline correctly?

My aim is to plot the bias-variance decomposition of a cubic smoothing spline for varying degrees of freedom. First I simulate a test-set (matrix) and a train-set (matrix). Then I iterate over 100 simulations and vary in each iteration the degrees…