I have deployed my running application in AKS. I want to add new disk (Harddisk of 30GB) but I don't know how to do it.
I want to attach 3 disks.
Here is details of AKS:
Node size: Standard_DS2_v2
Node pools: 1 node pool
Storage is:
default (default) kubernetes.io/azure-disk Delete WaitForFirstConsumer true
Please, tell me how to add it.
Based on Kubernetes documentation:
A PersistentVolume (PV) is a piece of storage in the cluster that has been provisioned by an administrator or dynamically provisioned using Storage Classes.
It is a resource in the cluster just like a node is a cluster resource. PVs are volume plugins like Volumes, but have a lifecycle independent of any individual Pod that uses the PV.
In the Azure documentation one can find clear guides how to:
create a static volume using Azure Disks
create a static volume using Azure Files
create a dynamic volume using Azure Disks
create a dynamic volume using Azure Files
NOTE:
Before you begin you should have existing AKS cluster and Azure CLI version 2.0.59 or later installed and configured. To check your version run:
az --version
See also this documentation.
A persistent volume represents a piece of storage that has been provisioned for use with Kubernetes pods. A persistent volume can be used by one or many pods, and can be dynamically or statically provisioned.
This article shows you how to dynamically create persistent volumes with Azure disks for use by a single pod in an Azure Kubernetes Service (AKS) cluster.
But if your requirement is to share the persistent volume across the multiple nodes use Azure FileShare
This article shows you how to dynamically create an Azure Files share for use by multiple pods in an Azure Kubernetes Service (AKS) cluster.
Related
Here, Actually we are running the Application in AKS Cluster and store the Application related data we are using the PV&PVC concept we need the backup of the data and has to store in Azure container so is it possible It need a suggestions and best approaches.
You can backup and restore your Azure Kubernetes Service (AKS) cluster's persistent volumes in a variety of methods. I would suggest you to use Velero so that you will not loss anything, in case of unintentional deletion or other possible failures.
Please use this article Azure Kubernetes Services (AKS). Backup/Restore your AKS data with Velero by Andrej Trusevic follow accordingly.
Install velero
After successfully script execution Velero will be created in your cluster
Stateless application backup & restore once backup is completed you can be able to see your backup files in the storage account blob container
For Reference:
Secure and back up your data
When creating an HDInsight cluster in Azure it is possible to set a specific blob container inside the selected storage account, so it is possible to re-use the same container when creating a new cluster in a second moment.
When creating an on-demand cluster using the Data Factory's HDInsight action, I can only specify the storage account but not the blob container, so at every pipeline run a new container is generated.
Is there a way to avoid that behaviour and setting the container name for the on-demand HDInsght cluster?
First of all, we cannot specify the container when creating the linked service. We can only specify the container when creating the dataset.
Sharing one blob container as the default file system for multiple clusters isn't supported. You can refer this documentation.
Sharing one blob container as the default file system for multiple clusters isn't supported.
Not able to set up persistent volume using Azure disk
We are trying to deploy an application on AKS and the application is to use persistent volume. If we use Azure disk, we have noticed if the node having the pod running the application container is stopped / not working , another pod from another node is spinned up but it is no longer accessing the persistent volume.
As per documentation ,azure disk is mapped to a particular node and file share is shared across nodes. What is the way to ensure that a application running on AKS using persistent volume is not lost if a pod/node does not work ?
We are looking for a solution with regard to persistent storage so that an application with 3 pods as a replica set can use an Azure disk persistent volume in AKS.
The Azure disk to work as the persistent storage volume in AKS, it should associates to the actual node, so it cannot share the files between multiple pods. So if you want to share files and persist files between pods whenever the pods in any node, the Azure File Share is a good way for you.
Finally, all of all, if you have multiple nodes and the deployment has 3 replicas. Then the best way to share and persist data between pods is using the Azure File Share or the NFS.
I just installed a DC/OS Cluster on Azure using Terraform. Now I was wondering if it's possible to automatically mount Data Disks of agent nodes under /dcos/volume<N>. As far as I understood the docs, this is a manual task. Wouldn't it be possible to automate this step with Terraform? I was looking through the DC/OS docs and Terraform docs but I couldn't find anything related to auto mounting.
It seems you just can mount the Data disks to the node of AKS manual as a volume. It's a Kubernetes task, not Azure's. Azure only can manage the data disk for you.
What you can do through the Terraform is attach the data disk to the node itself of AKS as a disk, not a volume of the AKS. And the volume, you only can create it through Kubernetes, not Azure. So Terraform also cannot help you achieve it automated.
I want to attach an external disk to kubernetes pod in azure environment. According to documentation here https://github.com/kubernetes/kubernetes/tree/release-1.2/examples/azure_file it uses azure file system.
What if I want to use the OS disks (external disks) like we have in gcloud environment ?
The Azure volumes piece was merged: https://github.com/kubernetes/kubernetes/blob/release-1.2/examples/azure_file/README.md