How can I decompress my .zip file and store in ADL/Blob storage? - azure

I have a ftp as a source connection where some I have zip file and others are not in compress form. I want to copy the files from ftp,decompress zip files and put all files into azure data lake or azure blob storage wherever it's possible to get decompressed.
I'm using copy data activity where I have a source as ftp and properties is zipDeflate,fastest and binary copy and the sink side, I'm just defining the destination ADL path. The files are getting copied to ADL but they're copying in compress form only.
Please let me know if it's possible to achieve the above objective by using copy activity process?

Using binary copy is your issue here, data factory wont understand the data it is moving to uncompress it. Try the same setup without binary copy!
Hope this helped!

Related

Azure Data Factory Copy Behaviour "Preserve Hierarchy" not working

I am trying to copy data from one container in Azure Data lake Gen2 into another in the same Storage Account. I want preserve the same hierarchy with folders and subfolders but whatever I try it does only copy the json file and no folders.
As of now I have the target container set in the target dataset. Should I add something more (such as directory and file)?
I have tested this for you and it can work, please follow this:
1.My container's structure:
examplecontainer
+test
+re
json files
+pd
json files
Setting of Source in Copy activity:
3.Setting of Sink in Copy activity:
4.Result:

Get ZIP file from SFTP to Azure Datalake using Azure data factory and store it into CSV format

Hi Have CSV files in zipped format in SFTP
I am able to install IR for SFTP and connection is successfull, but copying ZIP file from SFTP source to store into ADLS i getting error(import schema failed or no format define).
I have used copy data tool and dataflow for this scenerio but i am not sure how to zopy zip file from SFTP inside ADF.
Can you please try below:
Change to setting "ZipDeflate" compression type in your source data set and in the sink data set of Copy activity you don't need to specify any compression configuration (Compression type is "none").
In the Copy activity sink settings, please set the copy behavior to "Flatten Hierarchy" to unzip and write the individual files.

Get Meta data in Azure Data Factory is not giving me my multiple excel files successfully

I am trying to process multiple excel files in ADF to utilize them in a copy data activity to blob storage. Here is how my hierarchy is structured :
My source is an excel sheet coming from SFTP server (linked service).
File path: unnamed folder with multiple .xlsx files. Inside those files, the sheet name varies between sheet1 and table1.
I am trying to create get metadata to get all those files to pass them into a copy activity, but my metadata is never succeeding
Attached below is an elaboration about the problem:
If you only want to copy all excel files from SFTP to Blob Storage, there is no need to use Get Metadata activity.
Please try like this:
1.create binary format dataset
2.choose Wildcard file path when copy data
3.sink to your Blob Storage.

How to create a zip file from Azure blob storage container files using Pipeline

I have some dynamically created files in a blob storage container. I want to send it through email as a single attachment.
The total file size is less than 5 MB.
But here the difficulty I am facing is, when I try to compress the file using CopyData options, the compressed/zipped file not creating properly with multiple files.
If I try to zip a single file by giving its full path and filename, it is working fine. But when I give a folder name to compress all the files in that folder, it is not working correctly.
Please note that here I am not using any kind of external C# code or libraries.
Any help appreciated
Thank you
You can reference my settings in Data Factory Copy active:
Source settings:
Source dataset settings:
Sink settings:
Sink dataset settings:
Pipeline works ok:
Check the zip file in contianer containerleon:
Hope this helps.

Azure Data Lake .Net SDK Concat Files without removing files

Is there any way to concat files in data lake store with .net sdk for azure data lake store, that not delete from data lake files that i want to concat (merge) in one file? I'm searched in documentation but currently I couldn't find any method that not remove files after concat it in one file.
Data Lake Store does not support such an operation today. Concatenating files while keeping the source files intact amounts to copying the data, which is not currently supported as a built-in operation by Data Lake Store.
If you'd like to concatenate files and keep the source files intact, I recommend running the ADLCopy client to copy the source files into a separate/new folder, then run the Concat or MsConcat operation to destructively concatenate the new copy of the source files, not touching the original source files.

Resources