I have a large number of text files I need to grep through on a regular basis.
There are ~230,000 files amounting to around 15GB of data.
I've read the following threads:
- How to use grep efficiently?
- How to use grep with large (millions) number of files to search for string and get result in few minutes
The machine I'll be grepping on is an Intel Core i3 (i.e. dual-core), so I can't parallelize to any great extent. The machine is running Ubuntu and I'd prefer to do everything via the command line.
Instead of running a bog-standard grep each time, is there any way I can either index or tag the contents of the text files to improve searching?