Disaster recovery guidance for data in Data Lake Store
The data in your Azure Data Lake Store account is resilient to transient hardware failures within a region through automated replicas. This ensures durability and high availability, meeting the Azure Data Lake Store SLA. This article provides guidance on how to further protect your data from rare region-wide outages or accidental deletions.
Disaster recovery guidance
It is critical for every customer to prepare their own disaster recovery plan. Please refer to the Azure documentation below to build your disaster recovery plan. Here are some resources that can help you create your own plan.
We recommend that you copy your critical data to another Data Lake Store account in another region with a frequency aligned to the needs of your disaster recovery plan. There are a variety of methods to copy data including ADLCopy, Azure PowerShell or Azure Data Factory. Azure Data Factory is a useful service for creating and deploying data movement pipelines on a recurring basis.
If a regional outage occurs, you can then access your data in the region where the data was copied.You can monitor the Azure Service Health Dashboard to determine the Azure service status across the globe.
Data corruption or accidental deletion recovery guidance
While Azure Data Lake Store provides data resiliency through automated replicas, this does not prevent your application (or developers/users) from corrupting data or accidentally deleting it.
To prevent accidental deletion, we recommend that you first set the correct access policies for your Data Lake Store account. This includes applying Azure resource locks to lock down important resources as well as applying account and file level access control using the available Data Lake Store security features. We also recommend that you routinely create copies of your critical data using ADLCopy, Azure PowerShell or Azure Data Factory in another Data Lake Store account, folder, or Azure subscription. This can be used to recover from a data corruption or deletion incident. Azure Data Factory is a useful service for creating and deploying data movement pipelines on a recurring basis.
Organizations can also enable diagnostic logging for their Azure Data Lake Store account to collect data access audit trails that provides information about who might have deleted or updated a file.