I want to create a linux VM that can accomodate up to 10 terabytes of data. Not sure how to accomplish that on Microsoft Azure or if it is even possible. Any insight would be appreciated.
You create the storage separately from the VM - they have a data servuce known as BLOB service in Azure
Here is a link http://azure.microsoft.com/en-us/documentation/articles/storage-dotnet-how-to-use-blobs/
supporting many terrabytes is not a problem
You create one Linux VM from image and then attach (via Azure portal, like here) 10 disks of 1TB each. Today 1TB is the max size of a disk in Azure. As for the VM, you will need to have an Extra Large VM in order to accept this number of disks (up to 16 disks for an XL).
Related
I have two VMs using the same vnet and I would like to be able to copy a directory that has thousands of files and is about 400mb.
I can use a UNC path to copy the files, but that takes 2 minutes. I’ve also tried using a storage account and created a file share, but that is also slow.
Are there any other Azure resources that might make getting files from one VM to another faster?
As the comment point it out, if you have two VMs in the same VNet, you should use its private address. Traffic between the two VMs is kept on the Azure backbone network. You could directly copy/paste the files from one VM to another VM when you RDP to that VM.
Also, different VM size has different performance. For the best performance, It's recommended that you migrate any VM disk that requires high IOPS to Premium Storage. VM disks that use Premium Storage store data on solid-state drives (SSDs).
High-performance Premium Storage and managed disks for VMs
I have an Azure Windows Server VM running for several months.
Because of some historical reasons, I have two almost empty volumes on my Windows Server, I want to delete them to get rid of expensive bills.
I'm looking into ways to Delete volumes from Azure Windows Sever, and I believe you can achieve this by delete resources from Azure Portal.
However, I'm struggling finding the Mapping between DataDisk Resource and Windows Volume. (from my research, I think there likely to be an one-to-one mapping)
And for certain reason, I'm saving starting a new VM and migrate everything as the last resort.
Thanks!
UPDATE with details:
As I'm really struggling with this AZURE structure, I'm updating it with a lot of screenshots:
this is I have from all resources:
You can see I have 2 1TB disks, one is premium and another is standard
this is I have from clicking my virtual machine then click disks:
BUT IN HERE, I END UP HAVE TWO PREMIUM DISKS
And the following is powershell output on my Windows Server (disk4 does not have a LUN in the output):
I MUST BE MISSING SOMETHING, THINGS JUST DONT ADD UP!
I'm struggling finding the Mapping between DataDisk Resource and
Windows Volume.
You are right, Azure data disk and windows Volume to be an one-to-one mapping.
We can use get-disk to find the windows volume information, in this way, we can find LUN of this volume.
Here is my test, Windows server 2016.
Run get-disk command in Azure VM:
Via Azure portal we can find the LUN:
In this way, we can find which Azure VM disk map to windows volume.
Update:
Module: storage space, this means that disk create from storage pools(storage space).
Like this:
In this scenario, we can use this command Get-PhysicalDisk to get the disk information:
Also we can find the storage pool in Azure VM:
Update2:
Q1: Yes, we map Physical disk to storage pool, the create one or more disks in from this storage pool. For example, Physical disk about 1TB, we map this disk to storage pool, so the storage pool is 1TB, then we can create a disk from this storage pool, this disk should be smaller than 1TB, if we create 3 disks, disk1 + disk2 + disk3 = 1TB.
Q2: because he only create a disk from this storage pool, and the disk = 1TB, that means the disk = the storage space = TB, so the free space is 0.00B
Q3: yes, Unattached means this disk not in use at that time.
I want to use the premium storage for better performance.
I am using it for BLOBS and i need the fastest blob access for reading.
I am using the reading and writing of the blobs only internally within the data center
I create a premium storage and checked it vs the standard storage by reading a blob of 10 MB 100 times in different location using seek method (reading 50 kb each time).
I read it using a VM machine with windows server 2012
the result are the same - around 200 ms.
Do i need to do something else ? like attach the storage ? if so how do i attach the storage.
both the vm and the storage are at the same region
You can use Premium Storage blobs directly via the REST API. Performance will be better that Standard Storage blobs. Perf difference may not be obvious in some cases if there is local caching on the application or when the blob is too small. Here 10MB blob size is tiny compared to the performance limits. Can you retry with a larger blob? Like, 10 GB? Also note that Premium Storage model is not optimized for tiny blobs.
Well, in Virtual machine cases it always rely on your main Physical HDD, unless you will used that premium storage it's plus but i think internet connection matters as well.
By default, there is a temporary storage(SSD) provided with each VM. This temporary storage drive is present on the physical machine which is hosting your VM and hence can have higher IOPS and lower latency when compared to the persistent storage like data disk.
For test, we can create a VM with HDD disk, and attach a SSD to this VM. After it complete, we can install some tools to measure disk performance, in this way, we can find the difference between HDD and SSD.
like attach the storage ? if so how do i attach the storage.
We can via Azure new portal to attach a SSD to this VM.
More information about attach disk to VM, please refer to this link.
When creating the VM I'm asked about Storage configuration. When I select IOPS=0 (the minimum is otherwise 5000), Throughput=0 and Storage size=0, the info text is
0 data disks will be added to the virtual machine. This value was computed based on the value of IOPS, throughput, and storage size.
When the VM is created and I go to the Storage account, select Blobs and Container named vhds I see two disks, one 127GB and one 1TB disk.
Since the 1TB premium disks costs >100€/month I don't want that.
I tried removing the disk from a created machine but when I tried to add a new I got the error that "LUN :0 is already in use".
Preferably I would like to create machine correctly from the start. How can I do that?
This is correct. The current SQL Server IaaS experience on Azure Portal would creates one disk of 1TB even if specify 0 IOPS. We will add a fix to ensure the user cannot specify IOPS below 1 TB disk. If you need SQLVM without disks or any other configurations, you may use Azure PowerShell to create the VM.
I have created a VM in hyperv using dynamically expanding VHD option.Now I have exported that VHD to azure cloud and created a VM in azure.My VM is not booting properly. Is it recommended to use dynamically expanding VHD to create VM in azure?
OS and Data disks (ephemeral storage is local to the hardware) are thin provisioned in Azure. That is they are dynamically expanding. This is a decision made by Azure (so they don't have to store hundreds of thousands of mostly empty 120gb C: drives, which also makes it cheaper for you. ) and there is no option to change that.
Disks that you upload need to be a fixed size. (so you will pay for the full size)
According to Azure Documentation - Upload a Windows VM image to Microsoft Azure for Resource Manager deployments
Microsoft Azure can only accept images for Generation 1 virtual machines saved in the VHD file format. The VHD size must be fixed and a whole number of MB. The maximum size allowed for the VHD is 1023GB.
If you are having problems with an uploaded disk, it is likely that the problem is that you have created a dynamically expanding one and it simply isn't booting.
Usually, you want to create the VM with a fixed size disk as the dynamically expanding disks have a performance cost.
More information here:
https://technet.microsoft.com/en-us/magazine/ff458359.aspx