We're using a copy activity to copy parquet file data into our managed instance SQL server.
The source is using a SQL Serverless query to read the parquet files. There's a new column coming through that is bringing in large values and causing failures e.g. 28557632721941551956925858310928928
There isn't any problem querying it straight out of Azure Data Studio using SQL Serverless.
Here's the error message:
{ "errorCode": "2200", "message": "Failure happened on 'Source' side. ErrorCode=UserErrorInvalidDataValue,'Type=Microsoft.DataTransfer.Common.Shared.HybridDeliveryException,Message=Failed to read data from source.,Source=Microsoft.DataTransfer.ClientLibrary,''Type=System.OverflowException,Message=Conversion overflows.,Source=System.Data,'", "failureType": "UserError", "target": "Stage Parquet File Data", "details": [] }
I also tried using a parquet file dataset for my source. This is the failure I received:
{ "errorCode": "2200", "message": "ErrorCode=ParquetBridgeInvalidData,'Type=Microsoft.DataTransfer.Common.Shared.HybridDeliveryException,Message=Column gwcbi___seqval of primitive type FixedLenByteArray, original type Decimal contained an invalid value for the given original type.,Source=Microsoft.DataTransfer.Richfile.ParquetTransferPlugin,'", "failureType": "UserError", "target": "Stage Parquet File Data", "details": [] }
This looks like a serious limitation of Synapse/ADF pipelines. Any ideas?
Thanks, Jason