dataflow sink to update the source file

arkiboys 9,641 Reputation points
2022-04-26T12:18:29.177+00:00

dataflow1 has the following:
source1 --> aggregate --> sink1

source1 --> dsDatacompanies
sink1 --> dsDatacompanies

Note that source reads a .csv
aggregate then gets the distinct rows
sink1 then writes to the same file as source.

Is this ok or should the sink file be different to that of source?

Thank you

Azure Data Factory
Azure Data Factory
An Azure service for ingesting, preparing, and transforming data at scale.
9,601 questions
0 comments No comments
{count} votes

Accepted answer
  1. Nasreen Akter 10,751 Reputation points
    2022-04-26T14:01:56.557+00:00

    Hi @arkiboys ,

    I would create a separate output file than overwriting the source file. Note that you can use the same dataset as SOURCE and as SINK, only you have to parameterize the dataset e.g., if you want to create the file in a different folder, param the folder path, if you want to create the file in the same folder with different name --> param the filename. Thanks!

    0 comments No comments

0 additional answers

Sort by: Most helpful