I have a setup where there are multiple Apache / PHP servers collecting data from a request - Mainly the Get parameters and doing processing on it and saving it to either a database or flat file. The database is fine as everyone can connect independently and do the updates but with flat file, I am using a custom made queue which sends all data to one server where the flat file resides. My questions are:
Are there any good and reliable Log files processing systems I can use. I basically need to aggregate data coming in the log files and save it after some post processing into DB. If this is possible, I can simply have a nginx to log all requests (access.log) and run backend daemons to crunch the logs. I receive around 1000+ requests / second so I definately need a very robust system
Are there any good queue systems compatible with PHP and which is shared across multiple machines. A solution mainly over memcache where all information can be added from any node and can be accessed at any node at a very high speed. I need a system which can take bulk of data form the queue every second, process it and do the needful with it (save in DB). I dont think having queues on individual servers is very scalable as What I need to do is some level of aggregation before saving data. Since data can come on any of the PHP server (in round robin method), I do processing in MySQL (and hence use complicated queries), if I could collect all data at one server and let it do some processing and save in DB, that would ease my job
Thanks Sparsh