I have a work with azure etc,
I build a job in ssis that contains a for each for ADLS (Azure data lake store), it produces the path for some files. i make a data flow task inside the for each and add a ADLS Source with an expression for ADLS file path is dynamic by a variable that my for each produce.
When I run it, it always produces an error :
but when I write the variable that for each product to the file path it runs correctly (which is just 1 source, not all sources from for each)
Did anyone get an idea?
ADLS source does not support files at root folder.
If such a file is enumerated as passed to ADLS source, you get the error.
Related
I am trying to copy data from one container in Azure Data lake Gen2 into another in the same Storage Account. I want preserve the same hierarchy with folders and subfolders but whatever I try it does only copy the json file and no folders.
As of now I have the target container set in the target dataset. Should I add something more (such as directory and file)?
I have tested this for you and it can work, please follow this:
1.My container's structure:
examplecontainer
+test
+re
json files
+pd
json files
Setting of Source in Copy activity:
3.Setting of Sink in Copy activity:
4.Result:
Hi Have CSV files in zipped format in SFTP
I am able to install IR for SFTP and connection is successfull, but copying ZIP file from SFTP source to store into ADLS i getting error(import schema failed or no format define).
I have used copy data tool and dataflow for this scenerio but i am not sure how to zopy zip file from SFTP inside ADF.
Can you please try below:
Change to setting "ZipDeflate" compression type in your source data set and in the sink data set of Copy activity you don't need to specify any compression configuration (Compression type is "none").
In the Copy activity sink settings, please set the copy behavior to "Flatten Hierarchy" to unzip and write the individual files.
I am trying to process multiple excel files in ADF to utilize them in a copy data activity to blob storage. Here is how my hierarchy is structured :
My source is an excel sheet coming from SFTP server (linked service).
File path: unnamed folder with multiple .xlsx files. Inside those files, the sheet name varies between sheet1 and table1.
I am trying to create get metadata to get all those files to pass them into a copy activity, but my metadata is never succeeding
Attached below is an elaboration about the problem:
If you only want to copy all excel files from SFTP to Blob Storage, there is no need to use Get Metadata activity.
Please try like this:
1.create binary format dataset
2.choose Wildcard file path when copy data
3.sink to your Blob Storage.
I have some dynamically created files in a blob storage container. I want to send it through email as a single attachment.
The total file size is less than 5 MB.
But here the difficulty I am facing is, when I try to compress the file using CopyData options, the compressed/zipped file not creating properly with multiple files.
If I try to zip a single file by giving its full path and filename, it is working fine. But when I give a folder name to compress all the files in that folder, it is not working correctly.
Please note that here I am not using any kind of external C# code or libraries.
Any help appreciated
Thank you
You can reference my settings in Data Factory Copy active:
Source settings:
Source dataset settings:
Sink settings:
Sink dataset settings:
Pipeline works ok:
Check the zip file in contianer containerleon:
Hope this helps.
I have a ftp as a source connection where some I have zip file and others are not in compress form. I want to copy the files from ftp,decompress zip files and put all files into azure data lake or azure blob storage wherever it's possible to get decompressed.
I'm using copy data activity where I have a source as ftp and properties is zipDeflate,fastest and binary copy and the sink side, I'm just defining the destination ADL path. The files are getting copied to ADL but they're copying in compress form only.
Please let me know if it's possible to achieve the above objective by using copy activity process?
Using binary copy is your issue here, data factory wont understand the data it is moving to uncompress it. Try the same setup without binary copy!
Hope this helped!