how to replicate ongoing changes of on prem database to data lake?
Like AWS DMS handles ongoing replication and update bucket with respective files, what is available in azure to manage ongoing changes occurring at on-prem sql database? Initially it requires to do a full load in my data lake storage which can be…
Extract the date hierarchy from previous data in Data Lake Partition
Hi Team, Would be of great help to get some clear assistance on the below query. Data source is Dynamic365 & SAP Hana. My requirement is to create a data lake partition(ADLS Gen2) with date…
Can a new user assign to Posix style ACL's without RBAC in Data Lake storge gen2
Hi all, i have a active directory user say A i want to assign him Posix style ACL's- folder permission without RABC permission. it is possible or not please confirm
how can we create a custom role in Data Lake Store Gen 2 has access on azure storage user can view folder and container but not read file data
HI All, in Data Lake Gen2. we have one requirement- we need to create a custom role-adding user in this role User- should be able to view folder , but not able to read data. how can we create a custom role to get this requirement
Access Control Exception on setting data lake storage gen1 as output
Hi, I am receiving JSON data as input of stream analytics job from IOT hub and trying to write it in Data Lake Storage Gen-1. The stream analytics job has read, write and execute permissions and I have read permission to Data Lake storage. According to…
More convenient service to read avro files from Azure Data Lake Gen2
Hi, I have to read lots of avro files created by an Event Hub Capture in a Data Lake Gen2. Data must be filtered, processed and then applied to train a machine learning model. I'm considering Azure Databricks and the Azure Machine Learning service…
Creating datasets in Azure Machine Learning service from more than 100 paths
Hi, I need to create a dataset in Azure Machine Learning service from an Azure Data Lake Gen2 registered as a Datastore. Data in the lake are 1000's of avro files stored by an Event Hub Capture following the pattern…
HDInsight Cluster create a file with write mask
Hi, We are using Data Lake Gen 1 as the data store for HDInsight Cluster version 3.6. I wrote a simple spark code to write a file using saveas command of pyspark. The file is created in the DataLake with a write mask on. As a result. Any other user is…
How to push data from Azure Data lake to remote file server (network file folder)
Hello Experts, I need your guidance on how to push files from Azure Data lake to remote file server(network file folder). Our requirement is to send files from Azure Data Lake Gen2 to Remote File Server( Network File Folder), once the data pushed…
Merge two different Azure Data Lake Storage Account under one subscription.
Hi There, I have two Storage accounts the old ADLS Gen1 under one Management Group and subscription and another new ADLS Gen 2 under entirely different Management Group and another subscription. Now, the requirement is to have complete data (Gen1 +…
Copy Multiple files into ADL Gen2
I have a Data Factory pipeline that currently copies files daily from a Google Storage account down to an Azure Storage Blob ADL Gen2 enabled. Source several different files, File1, File2, File3 etc, all have a data range in the file name…
Azure File storage SAS TOKEN
I can upload files to ADLS Gen 2 blob storage with AzCopy through OAuth authorization, but I am unable to upload to file storage with the same. It is asking for SAS token. Please tell me the process of generating SAS token. Thanks. [Note: As we…
SSIS connection manager for ADLS Gen 2
Could we use SSIS with ADLS gen2. I am trying to write file to Azure data lake gen 2 using Azure Blob Upload task and Azure Data Lake Store File System Task, but i am unable to do it. Thanks. [Note: As we migrate from MSDN, this question has been…
Data Lake and Environments - Best Practice
Hello All, Is it a best practice to have one Big Data Lake for all the environments (Dev, Stage, QA and Prod) or have a Data Lake for Prod and another for Non-Prod ... etc.? If we chose to share a data lake across environments, then audit will play…