I have a pretty complex pipeline that does several machine learning algorithms on a given data. One of the algos uses GPU for processing. I'd like to somehow distinguish several nodes of my cluster with available GPU units so the algo would run only on those partitions. Is it possible with Spark?
I understand that it's more or less uncommon to have specific resources in a general computation environment, but this algo runs once in a while and it stays still whole other time - I'd like to use these nodes, too.