I have a large CSV file that I need to get into Redshift.
It has ~5 million rows. A couple issues:
1) The file's first 10 lines are gibberish that I want deleted/excluded
2) Whenever I try to upload csv files, I always get this weird glitch where it says that it loaded successfully, but 0 rows are uploaded. The way I resolve this is by re-formatting the columns in the CSV (formatting numbers as numbers, dates as custom date "YYYY-MM-DD") and then re-saving it. For whatever reason, it works after that.
Does anyone have similar issues and what is the best way to get this csv file into redshift?
UPDATE - so im using the COPY command and copying my csv file from S3 into a Redshift table. HOWEVER, i get an error and i look in stl_load_errors and the "colname" and the "raw_field_value" are not matching. it's really bizzare. for example, the "colname" will be like "name" but the "raw_field_value" will be a phone number. I double checked my csv file and this is NOT the case. any thoughts? it's not just 1 occurrence but a LOT, and several rows being mixed up.