question

nb-5535 avatar image
0 Votes"
nb-5535 asked PRADEEPCHEEKATLA-MSFT commented

How to write to Azure Table storage from an Azure Synapse Notebook

Context

We are performing some transformations on data using a Azure Synapse notebook (Pyspark) and in turn are trying to store that data inside Azure Table Storage.

We haven't been able to find any information (articles/ documentation) on storing data inside Azure Table storage using a synapse notebook directly.

Here is some documentation we did find but it involves a synapse pipeline


Questions

  1. Firstly, is it possible to store data directly to Azure Table storage using a synapse notebook? If so can someone link to the relevant docs/ articles so we know how to perform.

  2. Finally, if this is not possible could someone suggest the best work around/ alternative method to perform in order to get this done? Perhaps running the notebook as part of a pipeline and then having a copy data activity to store inside Azure Table storage?






azure-synapse-analyticsazure-table-storage
· 1
5 |1600 characters needed characters left characters exceeded

Up to 10 attachments (including images) can be used with a maximum of 3.0 MiB each and 30.0 MiB total.

Hello @nb-5535,

Following up to see if the below suggestion was helpful. And, if you have any further query do let us know.


  • Please don't forget to click on 130616-image.png or upvote 130671-image.png button whenever the information provided helps you.

0 Votes 0 ·

1 Answer

PRADEEPCHEEKATLA-MSFT avatar image
1 Vote"
PRADEEPCHEEKATLA-MSFT answered PRADEEPCHEEKATLA-MSFT commented

Hello @nb-5535,

Thanks for the question and using MS Q&A platform.

Currently Azure Storage - Tables are not supported. Only Azure blobs support the HDFS interface required by Hadoop & Spark.

Appreciate if you could share the feedback on our feedback channel. Which would be open for the user community to upvote & comment on. This allows our product teams to effectively prioritize your request against our existing feature backlog and gives insight into the potential impact of implementing the suggested feature.

Workaround: The best way to work with this is to pull the data into Azure Data Lake Gen2 storage account, and then run your notebook on the storage account using Azure Synapse or Databricks and then write it back to your Azure Table storage. Synapse Pipelines should allow this, as well as Azure Data Factory.

200822-image.png

Hope this will help. Please let us know if any further queries.


  • Please don't forget to click on 130616-image.png or upvote 130671-image.png button whenever the information provided helps you. Original posters help the community find answers faster by identifying the correct answer. Here is how

  • Want a reminder to come back and check responses? Here is how to subscribe to a notification

  • If you are interested in joining the VM program and help shape the future of Q&A: Here is how you can be part of Q&A Volunteer Moderators


image.png (62.8 KiB)
· 6
5 |1600 characters needed characters left characters exceeded

Up to 10 attachments (including images) can be used with a maximum of 3.0 MiB each and 30.0 MiB total.

Great thanks for this.

Just a clarifying question in your final paragraph when you say

"and then write it back to your Azure Table storage. Synapse Pipelines should allow this, as well as Azure Data Factory."

How can we write back to Azure Table storage after transforming via a synapse notebook using both synapse and data factory pipelines (can you provide more details)?

Last I recall data factory doesn't allow notebook activities, so what are the intermediate steps along with some potential activities to do the job within the pipeline itself?







0 Votes 0 ·

Hello @nb-5535,

The wording write it back meaning use copy activity in ADF/Synapse to copy data from ADLS gen2 account to Azure Table storage.

Yes, ADF doesn't allow notebook activities. As mentioned above you need to use Azure Databricks/Synapse Notebooks for transformations.

Hope this will help. Please let us know if any further queries.

0 Votes 0 ·
nb-5535 avatar image nb-5535 PRADEEPCHEEKATLA-MSFT ·

Oh ok, so the "end to end" flow you are suggesting as a workaround is:

Azure Data Lake gen2 --> Synapse Notebook --> Azure Data Lake gen2 --> Azure Table storage?

Where you start with some data in data lake, pull it in and transform it using the notebook and then write that back to data lake gen2 and then from there use a copy data activity to move the data from data lake to Azure Table storage?

0 Votes 0 ·
Show more comments

A Synapse notebook is a web interface for you to create files that contain live code, visualizations, and narrative text. Notebooks are a good place to validate ideas and use quick experiments to get insights from your data.

0 Votes 0 ·

Hello @nb-5535,

Just checking in to see if the above answer helped. If this answers your query, do click Accept Answer and Up-Vote for the same. And, if you have any further query do let us know.

0 Votes 0 ·