I have a script for collecting data for different social media hashtags. The script currently makes a bunch of sequential HTTP requests, formats the data into a Pandas data frame, and saves it to a csv. For very popular hashtags, it takes hours to run.
I need to run this program for 1000+ individual hashtags. To save time, I'd like to run many instances concurrently, say, 50-100 instances at a time, each collecting different hashtags.
Assuming I change the CSV portion to utilize a cloud storage service instead, what else do I need to do in order to accomplish what I'm describing? If I have a list of all the hashtags I need, how to I set up AWS lambda or Google Functions in order to execute these concurrently, so that 50-100 instances are always running until all the data is collected?