The case is that we have multiple servers (40+) to scrape one same URL at the same time (to make sure we have smallest legacy) and save the data into the database (MySQL).
and the problem now is that: the data is switch back and forth. for example, the content would be A <-> B <-> A <-> B <-> A in few seconds due to the crawler/database legacy.
is there a good way to prevent it? We're writing the crawler with Perl but any language would be fine for us since we can borrow the idea behind.
Any tip would be really appreciated. Redis? ZeroMQ?
Thanks