Azure Data Lake .Net SDK Concat Files without removing files - azure

Is there any way to concat files in data lake store with .net sdk for azure data lake store, that not delete from data lake files that i want to concat (merge) in one file? I'm searched in documentation but currently I couldn't find any method that not remove files after concat it in one file.

Data Lake Store does not support such an operation today. Concatenating files while keeping the source files intact amounts to copying the data, which is not currently supported as a built-in operation by Data Lake Store.
If you'd like to concatenate files and keep the source files intact, I recommend running the ADLCopy client to copy the source files into a separate/new folder, then run the Concat or MsConcat operation to destructively concatenate the new copy of the source files, not touching the original source files.

Related

Extract Only Specific Files with Azure Data Factory

I have to extract only "daily" files from a folder in my C: drive into azure data factory but there are "weekly" files that I don't want to extract. Also, I can't separate the two files in different onprem folders. I have to do this for a client but first I'm practicing on my own computer. Here is the onprem folder that I'm referring to. So the ultimate goal is to only transfer the "daily" files out of the folder and into azure data factory
As suggested by #Scott Mildenberger in comments as your files have similar naming convention you can use wildcard file path to filter files with name.
Sample data
Dataset settings
Source Setting
In file path type select wildcard file path and give daily* It will filter all the files from folder with files contains daily.
Output

Azure Data Factory Copy Behaviour "Preserve Hierarchy" not working

I am trying to copy data from one container in Azure Data lake Gen2 into another in the same Storage Account. I want preserve the same hierarchy with folders and subfolders but whatever I try it does only copy the json file and no folders.
As of now I have the target container set in the target dataset. Should I add something more (such as directory and file)?
I have tested this for you and it can work, please follow this:
1.My container's structure:
examplecontainer
+test
+re
json files
+pd
json files
Setting of Source in Copy activity:
3.Setting of Sink in Copy activity:
4.Result:

Get ZIP file from SFTP to Azure Datalake using Azure data factory and store it into CSV format

Hi Have CSV files in zipped format in SFTP
I am able to install IR for SFTP and connection is successfull, but copying ZIP file from SFTP source to store into ADLS i getting error(import schema failed or no format define).
I have used copy data tool and dataflow for this scenerio but i am not sure how to zopy zip file from SFTP inside ADF.
Can you please try below:
Change to setting "ZipDeflate" compression type in your source data set and in the sink data set of Copy activity you don't need to specify any compression configuration (Compression type is "none").
In the Copy activity sink settings, please set the copy behavior to "Flatten Hierarchy" to unzip and write the individual files.

Get Meta data in Azure Data Factory is not giving me my multiple excel files successfully

I am trying to process multiple excel files in ADF to utilize them in a copy data activity to blob storage. Here is how my hierarchy is structured :
My source is an excel sheet coming from SFTP server (linked service).
File path: unnamed folder with multiple .xlsx files. Inside those files, the sheet name varies between sheet1 and table1.
I am trying to create get metadata to get all those files to pass them into a copy activity, but my metadata is never succeeding
Attached below is an elaboration about the problem:
If you only want to copy all excel files from SFTP to Blob Storage, there is no need to use Get Metadata activity.
Please try like this:
1.create binary format dataset
2.choose Wildcard file path when copy data
3.sink to your Blob Storage.

How can I decompress my .zip file and store in ADL/Blob storage?

I have a ftp as a source connection where some I have zip file and others are not in compress form. I want to copy the files from ftp,decompress zip files and put all files into azure data lake or azure blob storage wherever it's possible to get decompressed.
I'm using copy data activity where I have a source as ftp and properties is zipDeflate,fastest and binary copy and the sink side, I'm just defining the destination ADL path. The files are getting copied to ADL but they're copying in compress form only.
Please let me know if it's possible to achieve the above objective by using copy activity process?
Using binary copy is your issue here, data factory wont understand the data it is moving to uncompress it. Try the same setup without binary copy!
Hope this helped!

Resources