Migrate Azure Storage account to a different subscription - azure

I have more than one Azure subscriptions, one for myself and others for clients. Can I change the subscription for one of my storage accounts so that it's associated to one of my other subscriptions?
Cannot find a way to do it through the Azure Management Portal. Perhaps it can be done with PowerShell?
Thanks!

I think you're best approach would be to contact Microsoft Windows Azure support to see if they can help.

I wanted to post some updated guidance in case anyone else runs into this issue. To move a storage account, you can do it easily for V1 or V2. You can move Classic Storage but its not easy and I would suggest upgrading it to the newer ARM based V1 or V2 storage account. This assume within the region. Moving to a subscription in another region is a different matter. This answer is probably very different that it would have been back in 2014.
Move Azure Storage account to another region
Move classic storage
Move V1 or V2 Storage
Move to another region
Migrate classic storage to ARM

Related

Azure Blob storage lifecycle management

I am currently using Azure Blobs to store data for a project. I want Azure to automatically delete old entries (data points) which are older then X number of days. I have found the following documentation:
https://learn.microsoft.com/en-us/azure/storage/blobs/storage-lifecycle-management-concepts?tabs=azure-portal
It essentially says that this can be done using lifecycle management and defining a new rule.
However, this documentation is over 6 months old and I cannot seem to find an option to select lifecycle management and define a new rule.
Has anyone else encountered this problem or know where I can access lifecycle management for an Azure Blob as of 2020?
Yes, this is a feature available today, I just confirmed on a storage account. You need to make sure you are using a V2 storage account, it will not be present on a v1, or blob only storage account.
I was experiencing the same issue, where the option for Life cycle management wasn't available but it was available on other storage accounts.
Check the performance/access tier. If it's set to Premium then its Life cycle management isn't available. Try creating a storage account with Standard.
If your using an arm template try Standard_RAGRS for the sku parameter.
screenshot of storage account in portal:

Underlying tables,blobs and queues not copied over to new subscription

Team,
When i migrated the storage account from one subscription to another using a migration tool(mig-az) it did not migrate the underlying tables, queues and blobs storage. Is that an inherent nature of migration?
is there a way to copy over the tables to the storage account in newer subscription? When doing the copy, would there be any downtime to the resources in the older subscription? Would the data in the older sub be lost?
Having a brief read of the Github page for Mig-Az, this seems to be a tool that just creates ARM templates from existing resources and allows you to apply them elsewhere. If this is the case, then it's only going to create a storage account, not any of the underlying resources or data, this is the nature of ARM templates.
If both your subscriptions are under the same Azure AD tenant then you can just move them using the Azure portal, there is an option in the resource group to move resources. See here for instructions.
If the subscriptions are under different tenants then you are going to need to manually move your data.

What kind of Azure Storage Account did I create?

I created a new Storage account in the Azure portal and choosed an existing Resource Group. It did not create a classic storage account but some kind of resource group-ish storage account that doesn´t have all the options a classic storage account has.
As an example, I could create the "files" folder through code, but I can´t use the code: "blockBlob.UploadFromStream(fileStream);", it gives me error 400 bad request. The same code works when I upload to a classic storage account.
What kind of storage account is seen in my image? Which is more correct, to create a classic storage account (blue icon in image) or the one I did (green/white/grey icon in image)?
First, I would suggest you have a look at David's reply in this thread to know the difference between new Azure storage account and classic Azure storage account.
it gives me error 400 bad request.
There are a lot of issues to caused 400 error. I would suggest your check your code to find out detailed issues. Please test use your code to create container (container name must meet the limitations) to see whether it will work. It is better if you could provide key code.
You created an Azure Resource Manager storage account of type Premium storage in region "North Europe" within your given resource group.
There is not really a right or wrong. As almost always it depends on your use-case.
I want to suggest these docs and samples for getting started with code addressing block blobs and Azure storage in general. Run and explore this code against the storage emulator and/or live storage accounts (classic/ARM standard/ARM premium). May be this helps finding a bug or a misconfiguration in your project.
Azure Blob Storage Samples for .NET
Azure Storage Samples
The issue here has nothing to do with Classic vs Resource Manager. It's related to the fact that the storage account is of type "Premium."
Premium storage accounts are exclusively used for Azure disks (attached durable storage), which are Page Blobs.
Premium storage doesn't support generic blobs/tables/queues. For that, you'd need a non-premium storage account.

Azure Cloud Service(classic) does not autoscale with new Storage Account

I deployed WorkerRole to Azure Cloud Service (classic) in new portal. With this, I also created Azure Storage account for queue.
Try to add AutoScale rule, the storage account is not listed. Tried to select Other Resource and put Resource Identifier of storage, there's no Metric name listed.
Is it by design that classic Cloud Service and new Storage account not working together?
Storage account data (e.g. blobs, queues, containers, tables) are accessible simply with account name + key. Any app can work with them.
However, to manage/enumerate available storage accounts, there are Classic-created and ARM-created accounts, each with different API's.
The original Azure Service Management (ASM) API doesn't know anything about ARM resources. There's a fairly good chance that, since you're deploying to a Classic cloud service, it's using ASM only and will not be able to enumerate ARM-created storage accounts.
If you create a Classic storage account (which has zero difference in functionality), you should be able to see it as an option for auto-scale.
I have a bit more details on the differences in this answer.
At this time, it is not possible to autoscale anything based on a new "v2" storage account. It has nothing to do with the fact that you are using the classic Azure Cloud Service. I am having the same issue with using Azure App Services. In the end, I just created a classic storage account to use for the autoscaling. There is no difference in how you interact with the different types of storage accounts.

Difference between new and classic storage accounts in Azure

Azure has Storage accounts and Storage accounts (classic) in the Azure Portal.
What are the differences between them? Is there any reason to migrate from a classic storage account to a new storage account?
Classic storage accounts are created using existing Service Management API's (the REST API stack that's been available for the past several years). The newer storage accounts are created with the new Azure Resource Manager (ARM) API's (which are also wrapped in PowerShell and CLI now). Ultimately they provide the same resources to your apps, but they're created/managed differently, and there are a few nuanced differences (such as the ability to tag resources that are created via ARM scripts).
You can't convert a classic storage account (or any classic resource) to a newer type. You don't really need to anyway, unless you're trying to mix resources from classic and new, such as adding ARM-based virtual machines to a classic-based virtual network, or spin up an ARM-based VM from a vhd image sitting in a classic storage account (and for that example, you could always just copy the vhd to a new storage account). Note that, for general storage usage (blobs/tables/queues), you just need the URI and the primary (or secondary) key. With those, you can access your storage resources from anywhere, from any VM/website/etc, regardless if you're accessing storage from classic or new virtual machines, for example.
Check out this link for a general list of differences between classic and new resources.
One advantage of the new over the classic storage accounts is Storage Service Encryption (SSE):
Q: I have an existing classic storage account. Can I enable SSE on it?
A: No, SSE is only supported on Resource Manager storage accounts.
Q: How can I encrypt data in my classic storage account?
A: You can create a new Resource Manager storage account and copy your
data using AzCopy from your existing classic storage account to your
newly created Resource Manager storage account.
There is now a way to migrate Classic resources to the new ARM model. I've done a few myself and it worked as expected. Here's a guide from Microsoft:
https://learn.microsoft.com/en-us/azure/virtual-machines/virtual-machines-windows-ps-migration-classic-resource-manager
In addition to #David Makogon's answer, the new Azure Storage offers reselling resources to sub-subscriptions.
This means that you are able to buy storage from Azure and sell it to your customers.
You can now migrate Classic Storage Accounts to ARM from within Azure.
Settings --> Migrate to ARM
With Azure
With Powershell

Resources