I get Bad Request every time i try to create an azure function - azure

I'm still new in Azure.
Today, I try to create an azure function instance
but, everytime I try to create one I get a Bad Request Error
{
"status": "Failed",
"error": {
"code": "BadRequest",
"message": "Cannot update the site '{project-name}' because it uses x64 worker process which is not allowed in the target compute mode.",
"details": [
{
"message": "Cannot update the site '{project-name}' because it uses x64 worker process which is not allowed in the target compute mode."
},
{
"code": "BadRequest"
},
{}
]
}
}
I have no idea what it does mean. So, can you give me an Idea what's wrong? and How I supposed to do to fix it?.

The Free/Shared tiers only support 32-bit application architecture. You'll need Basic or higher to use 64-bit.

Related

DataFactory Debug Error - BadRequest - factory is deleted

I am creating a pipeline that only runs a simple "wait ", just for testing, because I am trying to understand why my others pipeline are returning errors (the same error).
When I try to debug, it sends the following error:
{
"code": "BadRequest",
"message": "Operation could not be completed as factory is deleted",
"target": "pipeline/Teste_ParaApagar/runid/f0e412a9-21a2-4d0f-ab28-c0287a484326",
"details": null,
"error": null
}
I searched everywhere, I canĀ“t find answer. Can you please help?
I think this is not a common error.
According to the policy, I think you should check your account to see if your ADF was disabled.

Error when creating AKS cluster using a reference for the subnet ID

I'm getting an error when I try to deploy an AKS cluster using an ARM template, if the vnetSubnetId in the agentPoolProfiles property is a reference. I've used this exact template before without problems (on October 4th) but now I'm seeing an error with multiple different clusters, and when I do it either through a VSTS pipeline, or manually using PowerShell.
The property is set up like this:
"agentPoolProfiles": [
{
"name": "agentpool",
"count": "[parameters('agentCount')]",
"vmSize": "[parameters('agentVMSize')]",
"osType": "Linux",
"dnsPrefix": "[variables('agentsEndpointDNSNamePrefix')]",
"osDiskSizeGB": "[parameters('agentOsDiskSizeGB')]",
"vnetSubnetID": "[reference(concat('Microsoft.Network/virtualNetworks/', variables('vnetName'))).subnets[0].id]"
}
]
The variable 'vnetName' is based on an input parameter I use for the cluster name, and the vnet itself 100% exists, and is deployed as part of the same template.
If I try to deploy a new cluster I get the following error:
Message: {
"code": "InvalidParameter",
"message": "The value of parameter agentPoolProfile.vnetSubnetID is invalid.",
"target": "agentPoolProfile.vnetSubnetID"
}
If I try to re-deploy a cluster, with no changes to the template or input parameters since it last worked, I get the following error:
Message: {
"code": "PropertyChangeNotAllowed",
"message": "Changing property 'agentPoolProfile.vnetSubnetID' is not allowed.",
"target": "agentPoolProfile.vnetSubnetID"
}
Has something changed that means I can no longer get the vnet ID at runtime? Does it need to be passed in as a parameter now? If something has changed, is there anywhere I can find out the details?
Edit: Just to clarify, for re-deploying a cluster, I have checked and there are no new subnets, and I'm seeing the same behavior on 3 different clusters with different VNets.
Switching from reference() to resourceId() did fix the problem so has been marked as the answer, but I'm still no clearer on why reference() stopped working, will update that here as well if I figure it out.
I think what happened is subnets[0].id returns wrong (DIFFERENT) subnetId. and this is what the error points out. You cannot change the subnetId after deploying the cluster.
Probably somebody created a new subnet in the vnet. But I'd say that overall the approach is flawed. you should build the resourceId() function or just pass it as a parameter

Microsoft.Solutions/applicationDefinitions internal server error

I'm trying to create a Service catalog managed application definition and get the following error :{
"error": {
"code": "InternalServerError",
"message": "Encountered an internal server error. The tracking id is '5d5bf382-f101-48d7-bd89-72c72536a4a9'."
}
}
Is there anyway to find out what causes this? I did come across before. And believe it might be something to do with the template. I've tested the following using visual studio and the deployments works correctly. Just uploading the app.zip with the templates causes this error.
You can either look at the Azure Portal and go to Event Monitor and find error records and discern from the text there what went wrong or use Get-AzureRmLog -CorrelationId GUID_goes_here and look from there.

Problems with pulling OneDrive/SharePoint Permissions

I am currently writing a script in nodejs to pull all DriveItems from the MS Graph API, the idea is to grab the permissions of each DriveItem so that a record can be kept and presented to the user. Our authentication with Graph is app based.
I am running asynchronous requests to MS graph which obviously results in a lot of simultaneous requests. After a few seconds of executing the script, MS graph starts returning the following errors from the endpoint and doesn't seem to recover. The status codes are a 500 internal server error, 504 gateway timeout, and 404 not found.
The number of errors returned isn't always the same; one request returns 856 failed requests, another returns 900.
Here are the different responses received from endpoints which represent true drives and driveitems:
Endpoint: https://graph.microsoft.com/v1.0/drives/(drive_id)/items/(driveitem_id)/permissions
500 Internal server error:
{
"error": {
"code": "-2146232060, Microsoft.SharePoint.SPException",
"message": "Exception from HRESULT: 0x80131904",
"innerError": {
"request-id": "794863b8-bbf1-4866-9604-f8fcec4afabe",
"date": "2016-11-18T05:41:35"
}
}
}
504 Gateway timeout:
{
"error": {
"code": "UnknownError",
"message": "",
"innerError": {
"request-id": "e91c3037-10b6-4523-88d3-7fc31107b719",
"date": "2016-11-18T05:46:21"
}
}
}
404 Not found error:
{
"error": {
"code": "itemNotFound",
"message": "The resource could not be found.",
"innerError": {
"request-id": "b9a8286d-ce8c-4353-b8f0-818fe0f609ea",
"date": "2016-11-18T05:41:35"
}
}
}
This errors occur for a period of time and then everything seems to work fine again.
I can confirm that these endpoints work fine when I use postman to get an individual endpoint and the same token as the script - it returns valid data. Also if I make the script synchronous, it works fine but this is not an acceptable solution as it is far too slow to process the thousands of files I am dealing with.
The only thing I can think of that would be a problem is the sheer amount of simultaneous requests happening? However I am not receiving a 429 too many requests error back from msgraph before this starts to happen.
Any help is greatly appreciated.

Azure ML: Getting Error 503: NoMoreResources to any web service API even when I only make 1 request

Getting the following response even when I make one request (concurrency set to 200) to a web service.
{ status: 503, headers: '{"content-length":"174","content-type":"application/json; charset=utf-8","etag":"\"8ce068bf420a485c8096065ea3e4f436\"","server":"Microsoft-HTTPAPI/2.0","x-ms-request-id":"d5c56cdd-644f-48ba-ba2b-6eb444975e4c","date":"Mon, 15 Feb 2016 04:54:01 GMT","connection":"close"}', body: '{"error":{"code":"ServiceUnavailable","message":"Service is temporarily unavailable.","details":[{"code":"NoMoreResources","message":"No resources available for request."}]}}' }
The request-response web service is a recommender retraining web service with the training set containing close to 200k records. The training set is already present in my ML studio dataset, only 10-15 extra records are passed in the request. The same experiment was working flawlessly till 13th Feb 2016. I have already tried increasing the concurrency but still the same issue. I even reduced the size of the training set to 20 records, still didn't work.
I have two web service both doing something similar and both aren't working since 13th Feb 2016.
Finally, I created a really small experiment ( skill.csv --> split row ---> web output ) which doesn't take any input. It just has to return some part of the dataset. Did not work, response code 503.
The logs I got are as follows
{
"version": "2014-10-01",
"diagnostics": [{
.....
{
"type": "GetResourceEndEvent",
"timestamp": 13.1362,
"resourceId": "5e2d653c2b214e4dad2927210af4a436.865467b9e7c5410e9ebe829abd0050cd.v1-default-111",
"status": "Failure",
"error": "The Uri for the target storage location is not specified. Please consider changing the request's location mode."
},
{
"type": "InitializationSummary",
"time": "2016-02-15T04:46:18.3651714Z",
"status": "Failure",
"error": "The Uri for the target storage location is not specified. Please consider changing the request's location mode."
}
]
}
What am I missing? Or am I doing it completely wrong?
Thank you in advance.
PS: Data is stored in mongoDB and then imported as CSV
This was an Azure problem. I quote the Microsoft guy,
We believe we have isolated the issue impacting tour service and we are currently working on a fix. We will be able to deploy this in the next couple of days. The problem is impacting only the ASIA AzureML region at this time, so if this is an option for you, might I suggest using a workspace in either the US or EU region until the fix gets rolled out here.
To view the complete discussion, click here

Resources