I need to import data from thousands of URLs, here is an example of the data:
[{"date":"20201006T120000Z","uri":"secret","val":"1765.756"},{"date":"20201006T120500Z","uri":"secret","val":"2015.09258"},{"date":"20201006T121000Z","uri":"secret","val":"2283.0885"}]
Since COPY doesn't support JSON format, i've been using this to import the data from some of the URLs:
CREATE TEMP TABLE stage(x jsonb);
COPY stage FROM PROGRAM 'curl https://.....';
insert into test_table select f.* from stage,
jsonb_populate_recordset(null::test_table, x) f;
But it is inefficient since it creates a table for every import and it imports a single url at a time. I would like to know if it is possible (through a tool, script or command) to read a file with all the URLs and copy their data into the database.