Hello muntazir abbas,
Thank you for posting your query here!
Adding on to the previous response, since Azure Data Factory Data Flow does not directly support processing binary data, you would need to use a Copy Data Activity first to unzip the files and then call the Data Flow activity to process the desired JSON files.
You may create two folders at the storage container. One folder will have the zip folder files in compressed format and the sink folder will store the copied data.
- Create a Linked Service as a connection string to establish the connection from the source and target.
- Create a dataset as CSV to point to the source and the target.
(SOURCE DATASET)
Note: Mark the source compression type based on the compressed file.
(TARGET DATASET)
3. Select the copy activity to copy data from the source to the destination.
(AT THE SOURCE COPY ACTIVITY)
(AT THE SINK COPY ACTIVITY)
(COPY ACTIVITY)
(COPIED DATA AT THE SINK FOLDER)
Apart from this, you can create an Azure Function with blob trigger, and the function would need to handle the logic to unzip the zip file. This solution is dependent on your code.
Or you can create an Azure Logic App, with Blob trigger, then use connectors to unzip the zipped files.
Reference: https://www.frankysnotes.com/2019/02/how-to-unzip-automatically-your-files.html
And then configure the Data Flow Activity to process the unzipped JSON files from the folder where they were extracted by the previous Copy Data Activity.
Please let us know if you have any further queries. I’m happy to assist you further.
Please do not forget to "Accept the answer” and “up-vote” wherever the information provided helps you, this can be beneficial to other community members.