After reading this post here: Azure Machine Learning Request Response latency and the article mentioned in the comments I was wondering if this behavior is also true when a published webservice is called in batch mode. Especially since I have read somewhere (sorry, can't find the link at the moment) that the batch calls are not influenced by the "concurrent calls" config...
In our scenario we have a custom R module uploaded to our workspace which includes some libraries that are not available on aML by default. The module takes a dataset, trains a binary tree, creates some plots and encodes them in base64 before returning those as a dataset. Locally that does not take more than 5s. But in the aML webservice it takes approx. 90s and it seems that the runtime in batchmode does not improve when calling the service multiple times.
Additionally it would be nice to know for how long the containers, mentioned in the linked post, will stay warm.