3

I have a copy pipeline set up in Azure Data Factory, which copies everything in Cosmos DB daily to Azure Data Lake. When copying. there is a spike on RU/s. I donot want to increase Throughput.

Anything I can do to lowering the impact? e.g. can I set a limit to the copying pipeline?

enter image description here

Daolin
  • 614
  • 1
  • 16
  • 41

1 Answers1

1

As @David said in the comment, any interactions with Cosmos DB requires the consumption of Rus. RUs setting is an important indicator of fees and performance. More details, you could refer to this official article.

Basically, RU metrics will be shocked by the adf copy activity and throughput setting will not be automatically adjusted by cosmos db.

If you do want to adjust throughput setting temporally,you could execute http trigger azure function with azure function activity which is accessed at head and tail of copy activity. In that activity, adjust throughput settings appropriately with sdk or rest api.(Please refer to the case:Cosmos Db Throughput)

Jay Gong
  • 23,163
  • 2
  • 27
  • 32
  • I'm running a simple adf pipeline for storing data from data lake to cosmos db (sql api). After setting database throughput to Autopilot 4000 RU/s, the run took ~11 min and I see 207 throttling requests. On setting database throughput to Autopilot 20,000 RU/s, the run took ~7 min and I see 744 throttling requests. Why is that? Thank you! – user989988 May 06 '20 at 19:12