I uploading a 1.8Gb CSV file on Google Cloud Storage.
When I start the training from the Google API Explorer I get an error:
{
"error": {
"errors": [
{
"domain": "global",
"reason": "invalid",
"message": "Data size limit exceeded."
}
],
"code": 400,
"message": "Data size limit exceeded."
}
}
I'm confusing. From the FAQ I can read:
What training data does the Prediction API support?
Training data can be provided in one of three ways:
A CSV formatted training data file up to 2.5GB in size, loaded into Google Storage.
And from the pricing page:
Training:
$0.002/MB bulk trained (maximum size of each dataset: 250MB)
What is the difference from this 250MB and the 2.5GB ?