I am trying to implement AdaBoost algorithm, and have two questions.
1) At each iteration, the training data has to be re-sampled in accordance with a probability distribution. Should the size of re-sampled data set be the same as the one of original data set. 2) If I re-sample the training data set according to a probability distribution, it is quite possible that I can get multiple copies for a single data point. Should I keep all of these redundant copies while training the weak classifier at each iteration.