I have a Azure website that allows customers to upload documents. There are a lot of documenet (~200Gb so far).
I need a way to backup the documents to another location (Azure or other location), or have live replication to a server. Is there anything I can use that will do this?
Have you looked at Azure storage redundancy options? The geo redundant option might solve your replication need.
http://blogs.msdn.com/b/windowsazurestorage/archive/2013/12/04/introducing-read-access-geo-replicated-storage-ra-grs-for-windows-azure-storage.aspx
You can use Cloudberry. Here is the link to their website. The tool will not give you cloud storage, but will assist in backup process. It can back up either to the cloud or to the local storages.
Related
I'm trying to make an architecture for a data lake, I already generated my CSV, txt, and Avro files they are in an On-Premise machine and I want to upload them to Google Cloud Storage, but I see that I have to go through the public internet and I don't want to that.
What options do I have to make as safe as possible?
I was trying to make a Compute Engine environment to upload the files here through SFTP and then moved them to Cloud Storage, but that will make my cost to go up.
I think this the use case of Google VPC. The key benefit provided on the product description is:
For on-premises, you can share a connection between VPC and on-premises resources with all regions in a single VPC
If you look deeper into documentation you may find document Configuring Private Google Access for on-premises hosts. Some information maybe be found here as well. This might be a good option for you.
I want to create a backup for azure SQL databases that are not currently used to some low-cost storage account. I don't want to pay for those dbs as no operation are done on them. But I might need that data in the future. Extraction time is not an issue. So just want to know what are the different methods to do the same.
If I understood the question correctly then I would suggest for that database which you might not need immediately you can take back on azure blob storage.
There is a good blog for it
https://dallasdbas.com/sql-backups-azure-storage/
Just follow it.
Hope this will help you.
I am very new to Azure, so posting this query. I just want to copy a flat file from an on-prem server to Azure BLOB storage daily basis. I thought "azcopy" will be a good solution for this requirement. But challenge is we can not configure "azcopy" utility in that on-prem server and can not schedule any cron job as this sever is owing by our client. So we thought of running "azcopy" utility from one of the Unix VM in Azure cloud and want to pull the file and transfer to Azure BLOB storage.
I am not getting any clear idea from documentation that "azcopy" can work as a pull manner or not.
Can anyone help me to understand whether my approach will work or not ? and if not then please give me some idea how to do this.
Please see below diagram depicting what I want to achieve.
AzCopy is not designed to be used in that way.
If there's any way for you to access the file, that will be up to your customer and you'd have to ask their network security folks how they would expect it to be done. But most customers (or at least all the ones I worked with before joining Microsoft) prefer push not pull models, since the security for push is more standard and easier to set up.
I want to spike whether azure and the cloud is a good fit for us.
We have a website where users upload documents to our currently hosted website.
Every document has an equivalent record in a database.
I am using terraform to create the azure infrastructure.
What is my best way of migrating the documents from the local file path on the server to azure?
Should I be using file storage or blob storage. I am confused about the difference.
Is there anything in terraform that can help with this?
Based on your comments, I would recommend storing them in Blob Storage. This service is suited for storing and serving unstructured data like files and images. There are many other features like redundancy, archiving etc. that you may find useful in your scenario.
File Storage is more suitable in Lift-and-Shift kind of scenarios where you're moving an on-prem application to the cloud and the application writes data to either local or network attached disk.
You may also find this article useful: https://learn.microsoft.com/en-us/azure/storage/common/storage-decide-blobs-files-disks
UPDATE
Regarding uploading files from local computer to Azure Storage, there are actually many options available:
Use a Storage Explorer like Microsoft's Storage Explorer.
Use AzCopy command-line tool.
Use Azure PowerShell Cmdlets.
Use Azure CLI.
Write your own code using any available Storage Client libraries or directly consuming REST API.
I have an account on azure and build a cluster on it (Chemalivethermotest). It has a storage account and I just by mistake erase all data on it (I misanderstood something on the online azure interface). Is there a way to recover the datas? I see there are no files anymore but there are still blob. Are they allowing to recover things?
Thanks you very much for your help.
This is an issue you will have to take up directly with Azure support, there is no technical answer. Good luck!