I'm trying to use Flume to ship my access logs to a Spark cluster. But there're a bunch of limitations that forces me to write a custom application (or a Flume source) to read the log files.
What I'm trying to do is to get Flume client to signal this source in case it can't write the data to the sink. As we constantly have long network outages and there's not enough disk space to queue the failed logs on disk until the network is back up. Instead I would like to "tell" the source to stop reading the logs until the network is up, then "tell" it to start again. But so far I haven't seen any kind of callbacks on errors in the documentation.
Is there anyway I can achieve such scenario without reinventing the wheel?