Thanks for your inputs! Please find my inline comments.
- Memory Management
- Increase Integration Runtime Capacity: We don't have control on this setting and we might not be given a chance to make any changes with the current situation.
- Optimize Data Volume: Yes, this is already tried and for which it works but takes hours and hours to get the pipeline completion.
- Buffer Size Configuration
- Adjust Buffer Settings: This was tried too but nothing has favored to resolve the issue.
- Use Staging: Yes, this is considered but there are some issues to be used this way which we are still evaluating. But will keep progressing on this as well.
- Entity-Specific Issues
- Analyze Entity Tables: not all have this kind of columns with large text but there are few entities that have this large text fields. Even without that it was able to completed 1 run and the subsequent run is failing where it is not giving a pattern on why it got succeeded in the first run but failed in the next runs.
- Data Transformation: No transformations are in scope for the current design. Also, to use a SP or Dataflow or Databricks we have some limitations when it comes to Odata as source.
- Error Handling and Logging
- Implement Retry Logic: Unless used the loop conditions and with the retry settings as well it is not getting succeeded.
- Enable Detailed Logging: This setting I am still trying to enable as the activity log entries are not showing any specific information.
- Review ADF Limits
- Check ADF Limits: To some extent we were able to verify and reduce the load accordingly. But if there is any information on the same please share.