I have observed in many articles and books that model selection is done before model tuning.
Model selection is generally done using some form of cross validation like k-fold where multiple models' metrics are calculated and the best one is selected.
And then the model selected is tuned to get the best hyperparameters.
But my question is that a model that was not selected might perform better with the right hyperparameters.
So why aren't all the models we are interested in tuned to get the right hyperparameters and then the best model be selected by cross validation.