I'm need to write a huge quantity of entities (1.5 million lines from a .csv file) to Google Cloud Datastore. Kind of a 2 part question:
Can I do (or is kind
a necessary property?):
const item = {
family: "chevrolet",
series: "impala",
data: {
sku: "chev-impala",
description: "Chevrolet Impala Sedan",
price: "20000"
}
}
then, regarding importing I'm unsure of how this works. If I can't simply dump/upload/import a huge .json
file, I wanted to use Node.js. I would like each entity to have an autogenerated universal id. Is there an asynchronous means of writing? I have a node script that is piping out a few hundred enteties/records at a time and pausing awaiting write resolve
. ...which is what I'm looking for : a promise import.