Skip to main content

Linked Service in Azure Data Factory

 

Linked Service provides Azure data factory with basic connection information required to connect external source. There are multiple ways to create a Linked Service in Azure - Via Manage in Azure Data Factory UI, Power Shell, Azure Portal.

Data Store Linked Service can be used to configure connection setup for data stores such as relational databases, Azure blob, on-prem FTP servers, HDFS, and many more.

Compute environments supported in ADF can be configured using Compute Linked Service. Azure Databricks, Azure Batch, HD Insights, Azure ML, and Azure Data lake Analytics are the platforms supported as of today.

Parameterizing Linked Service

The ability to parameterize a linked service makes it an extremely powerful utility. In a DW/BI, It is a fairly common scenario to have multiple source data systems using a sing RDBMS system, Oracle for instance. Using Parameters, we can configure a single linked service to connect to multiple homogeneous data systems. Parameter values can be set at datasets using this parameterized Linked Service. There is also a provision to set a default value for a parameter. Below snippets show the parameter section available while creating a new linked service.




 




Comments

Popular posts from this blog

How to work with XML files in Databricks using Python

This article will walk you through the basic steps of accessing and reading XML files placed at the filestore using python code in the community edition databricks notebook. We will also explore a few important functions available in the Spark XML maven library. Think of this article as a stepping stone in the databricks community edition. Features and functionalities elaborated herein can be scaled at the enterprise level using Enterprise editions of databricks to design reusable file processing frameworks. Requirements We will be using the Spark-XML package from Maven. **Spark 3.0 or above is required on your cluster for working with XML files. This article uses Databricks Community edition, refer to  this  video tutorial from Pragmatic works for getting started with  Databricks Community Edition . Create your first cluster in seconds : The next step is to install the Spark-XML library on your cluster. The cluster needs to be in a running state to install this li...

ACLs in Azure Data Lake

Access Control Lists(ACLs) in azure are an extremely powerful toolset to provision granular levels of access in Azure Data Lake. Role-Based Access Control (RBAC) is best option to setup broader access levels however with ACLs you can reach the lowest possible grains as low as a file inside a blob container. Think of a scenario where you want to add more than 1 user to a folder inside a blob container and each one of them sees only their data - Possible with ACLs Prerequisites Azure Subscription Storage blob with hierarchical namespace enabled Reader Access on the storage object via RBAC How to setup ACLs in Azure Data Lake Like any other offering, Microsoft has a broad spectrum of tools/ways to setup ACLs, ranging from Azure Portal to writing python code . All the steps involved are available in Microsoft documentation, and in a very descriptive manner, therefore needless to rephrase again in this article. Instead, lets walk through some of the challenges one can c...

import in Python

Python allows you to reuse a code using imports. It can be constant values, general functions, formulas and what not. For instance you are writing a code to implement circle circumference formula, instead of hardcoding PI, Import Maths module from Python Standard Library  . This library has a massive list of facilities on offer. Learn about - ' from '