Imagine, you are a librarian and during time you have classified a bunch of text files (approx 100) with a general ambiguous keyword.
Every text file is actually a topic of keyword_meaning1 or a topic of keyword_meaning2.
Which unsupervised learning approach would you use, to split the text files into two groups?
What precision (in percentage) of correct classification can be achieved according to a number of text files?
Or can be somehow indicated in one group, that there is a need of a librarian to check certain files, because they may be classifed incorrectly?