I just created a Gradient Boosting model whose out-of-sample prediction is worse than the random forest. The MSE of GBM is 10% higher than the random forest. Below is my sample code. I am sure whether there is any wrong with it.
gbm1 <- gbm(as.formula(paste0(Y.idx ,'~', paste0(colnames(rf.tmp.train[c(-1,-2)],collapse=""))),
data=rf.tmp.train,distribution="gaussian",n.trees=3000,
shrinkage=0.001,interaction.depth=1,bag.fraction = 0.5,
train.fraction = 1,n.minobsinnode = 10, cv.folds = 10,
keep.data=TRUE, verbose=FALSE,n.cores=1)