I am working on a solution to send DynamoDB data to an S3 bucket for multiple teams to consume. I need 2 separate solutions:
- Initial data dumb of the table (~ 3 TB)
- Continuous updates of the add/modify/delete operations (but not the schema) also sent to the S3 bucket.
Based on the latest functionality available from AWS, what is the most affordable solution for this, in terms of cost and time? Is there a no-code solution that can easily be created for this and get it up and running fast?
I have considered the following:
- PITR backup for initial load to S3
- DynamoDB Streams and/or Kinesis streams/firehose for updates/adds/deletes
Would EMR or Glue be helpful here? What else should I be considering?