InvalidResourceName error while writing parquet file in ADLS Gen2 - azure

I am reading a csv file and writing the same in parquet format in ADLS Gen2 using ADF copy activity.
My source:
My Sink :
I am facing the below error :
Failure type
User configuration issue
Details
ErrorCode=AdlsGen2OperationFailed,'Type=Microsoft.DataTransfer.Common.Shared.HybridDeliveryException,Message=ADLS Gen2 operation failed for: Operation returned an invalid status code 'BadRequest'. Account: 'adlsedmadifpoc'. FileSystem: 'raw_area'. ErrorCode: 'InvalidResourceName'. Message: 'The specifed resource name contains invalid characters.'. RequestId: '70d7xbfd-6xxf-00ec-2c74-9axxxx000000'. TimeStamp: 'Thu, 26 Aug 2021 12:19:56 GMT'..,Source=Microsoft.DataTransfer.ClientLibrary,''Type=Microsoft.Azure.Storage.Data.Models.ErrorSchemaException,Message=Operation returned an invalid status code 'BadRequest',Source=Microsoft.DataTransfer.ClientLibrary,'
Any help appreciated. Thanks.

This is caused when there are characters introduced to the Container Name which do not follow the Azure storage container naming convention.
The best way to avoid this error is to copy it as it is from the Azure environment.
A container name must contain valid characters, conforming to the following naming rules:
All letters must be lowercase.
Names must be from 3 through 63 characters long.
Names must start or end with a letter or number.
Names can contain only letters, numbers, and the dash (-) character.
Every dash (-) character must be immediately preceded and followed by a letter or number; consecutive dashes are not permitted in container names.

Related

ADF copy activity failed while extracting data from DB2- Issue found for few records having special characters

I am doing a full extract from a table ABC. In copy activity, I have given a query as
select * from ABC
whereas I am facing issue for few rows (It has special characters - Japanese and Korean)
Error code 2200
Failure type User configuration issue
Details Failure happened on 'Source' side. ErrorCode=DB2DriverRunFailed,'Type=Microsoft.DataTransfer.Common.Shared.HybridDeliveryException,Message=Error thrown from driver. Sql code: '-343',Source=Microsoft.DataTransfer.ClientLibrary.Db2Connector,''Type=Microsoft.HostIntegration.DrdaClient.DrdaException,Message=HISMPCB0001 In BasePrimitiveConverter an exception has occurred. Exception description: Output buffer is smaller than required size 12 SQLSTATE=HY000 SQLCODE=-343,Source=Microsoft.HostIntegration.Drda.Requester,'
The character which is causing the issue is '轎ᆃ '
In the error description, it states that there is BasePrimitiveConverter exception that has occurred. The exception description indicates that the output buffer is smaller than the required size. So, please try converting the column to an acceptable type like graphic in db2. Refer to the following link to understand more.
https://bytes.com/topic/db2/answers/488983-storing-some-japanese-data
Referring to these links, I understand that this error might be due to the datatype of source column, or the encoding used. Try working with different encoding options available in your source dataset. Here is a similar problem with a different source but a similar problem of not being able to retrieve special characters.
https://learn.microsoft.com/en-us/answers/questions/467456/failure-happened-source-side-in-copy-activity-for.html

How to filter blobs in Azure SDK for Python

I want to search for blobs in my Azure blob storage according to a specific tag (like: .name, .creation_date, .size...)
My current way is returning all blobs from the container with MyContainerClient.list_blobs and searching for the corresponding tag afterwards. Since my container stores around 800000 blobs, this takes me around 20 min, which is not usable for a live view of the content.
But I also found another ContainerClient function: .find_blobs_by_tags(filter_expression: str) where I can search for a specific blob whose tags matches the specified condition.
In the Azure API they specified this filter_expression as: ""yourtagname"='firsttag'" , therefore I specified: ""name"='example.jpg'" or ""creation_date"='2021-07-04 09:35:19+00:00'"
Azure SDK Python - ContainerClient.find_blobs_by_tag
Unfortunately I always get an error:
azure.core.exceptions.HttpResponseError: Error parsing query at or near character position 1: unexpected 'creation_time'
RequestId:63bd850b-401e-005f-745e-400d5a000000
Time:2022-03-25T15:40:22.4156367Z
ErrorCode:InvalidQueryParameterValue
queryparametername:where
queryparametervalue:'creation_time'='0529121f-7676-46c7-8a52-424664774240/0529121f-7676-46c7-8a52-424664774240.json'
reason:This query parameter value is invalid.
Content: <?xml version="1.0" encoding="utf-8"?>
<Error><Code>InvalidQueryParameterValue</Code><Message>Error parsing query at or near character position 1: unexpected &apos;creation_time&apos;
RequestId:63bd850b-401e-005f-745e-400d5a000000
Time:2022-03-25T15:40:22.4156367Z</Message><QueryParameterName>where</QueryParameterName><QueryParameterValue>&apos;creation_time&apos;=&apos;0529121f-7676-46c7-8a52-424664774240/0529121f-7676-46c7-8a52-424664774240.json&apos;</QueryParameterValue><Reason>This query parameter value is invalid.</Reason></Error>
Has someone experience with this Azure function calls?
Looking at the github code(in the find_blobs_by_tags function) , it says :
:param str filter_expression:
The expression to find blobs whose tags matches the specified condition.
eg. "\"yourtagname\"='firsttag' and \"yourtagname2\"='secondtag'"
Looks like you are missing the escape characters? Can you try including them in?

Video file not recognized in android studio

I added a file named 'testingClip' of mp4 format in a directory i made named 'raw' under 'res',it keeps on showing this error message saying it doesn't recognize the file type.
Don't use any capital letter in resource file name Change your testingClip.mp4 to testingclip.mp4
resource names must contain only lowercase a-z, 0-9, or underscore and The resource name must start with a letter.

spark executor pods creation failed

Im running spark 2.3.1 on kubernetes 1.11.0
Im getting below error when spark driver pod is trying to bringup executor pods ,it truncates first 7 letters in the pod name and throws error that the name starts with "-"
io.fabric8.kubernetes.client.KubernetesClientException: Failure
executing: POST at:
https://kubernetes.default.svc/api/v1/namespaces/mynamespace/pods.
Message: Pod
"mybrand-sb1-ca-privacy-abc469-38957af1c3393cae8941b0613376040c-exec-29"
is invalid: spec.hostname: Invalid value:
"-sb1-ca-privacy-abc469-38957af1c3393cae8941b0613376040c-exec-29": a
DNS-1123 label must consist of lower case alphanumeric characters or
'-', and must start and end with an alphanumeric character (e.g.
'my-name', or '123-abc', regex used for validation is
'a-z0-9?').
the max size of the name is 63 characters. Thats the reason its getting truncated for the first 9 characters. Try to minimize the app name, this will solve the issue.

Creating class from xsd with many imports

So I've ready alot of xsd documentation and similar questions but I've run into a problem: The specified path, file name, or both are too long. The fully qualified file name must be less than 260 characters, and the directory name must be less than 248 characters.
this is the command:
xsd BulkmailAnnouncement2.xsd Components1.xsd Components2.xsd Components3.xsd Components4.xsd Fields1.xsd ManufacturingComponents1.xsd ManufacturingComponents2.xsd Meta1.xsd QualifiedDataTypes.xsd UnqualifiedDataTypes.xsd CodeLists1.xsd CodeList_UnitCode_UNECE_7_04.xsd CodeList_MIMEMediaTypeCode_IANA_7_04.xsd CodeList_LanguageCode_ISO_7_04.xsd CodeList_CurrencyCode_ISO_7_04.xsd /c
How can I create a class if xsd cannot support the includes.
Move files to another directory or rename folders / files in the current path.
The problem occurs because the full path to the file / files exceeds the allowed limit.
Quote from the official documentation:
In the Windows API (with some exceptions discussed in the following
paragraphs), the maximum length for a path is MAX_PATH, which is
defined as 260 characters. A local path is structured in the following
order: drive letter, colon, backslash, name components separated by
backslashes, and a terminating null character. For example, the
maximum path on drive D is "D:\some 256-character path string"
where "" represents the invisible terminating null character for
the current system codepage. (The characters < > are used here for
visual clarity and cannot be part of a valid path string.)

Resources