I have a copy activity build in Azure Data Factory V2, where the datasource is a SFTP folder with several XML files and the Sink is a Azure Postgres Database.
I have successfully used the copy activity for small files (20 MB). But I have 3 major XML files with 3 GB, 4.5 GB and 18 GB.
- For the size of these files which settings should I choose? How many DIU?
- Is the choice of the datasource relevant? This is, using Amazon S3 or Data Blob is better that FTP? (I ask this because is it taking too long for just coping the data).