Azure Data Lake Storage Connection Details
Introduction
Connector Version
This documentation is based on version 25.0.9368 of the connector.
Get Started
Azure Data Lake Storage Version Support
The connector leverages the Azure Data Lake Storage API to enable bidirectional access to Azure Data Lake Storage.
Stream Azure Data Lake Storage Data
The Azure Data Lake Storage connector is designed for navigating the Azure Data Lake Storage metadata only. A variety of of stored procedures relevant to Azure Data Lake Storage data are supported as well.
This metadata typically includes details about stored objects, such as file and folder names, and excludes the actual content of the discoverable files.
If access to both the file metadata and the actual file content is needed, then the Azure Data Lake Storage connector must be used in tandem with the associated file streaming driver(s) for the filetypes stored in Azure Data Lake Storage.
The following file streaming drivers are available:
- Avro
- CSV
- Excel
- JSON
- Parquet
- REST
- XML
See the relevant file streaming driver's documentation for a configuration guide for connecting to files stored in Azure Data Lake Storage.
Establish a Connection
Connect to Azure Data Lake Storage Gen 2
To connect to a Gen 2 DataLakeStorage account, set the following properties:
- Account: The name of the storage account.
- FileSystem: The file system name used for this account. For example, the name of an Azure Blob Container.
- Directory (Optional): The path to the location where the replicated file should be stored. If no path is specified, the file is stored in the root directory.
Authenticate to Azure Data Lake Storage Gen 2
Azure Data Lake Storage Gen 2 supports five different ways to authenticate: using an Access key (AccessKey), using a Shared Access Signature (SAS), via Microsoft Entra ID OAuth (Microsoft Entra ID), via Microsoft Entra service principal (AzureServicePrincipal or AzureServicePrincipalCert), or via Managed Service Identity (AzureMSI).
Access Key
To connect using an access key, you must first obtain an available access key for the ADLS Gen2 storage account.
At the Azure portal:
- Go to your ADLS Gen2 Storage Account.
- Under
Settings, selectAccess keys. - Copy the value for one of the available access keys to the AccessKey connection property.
When you are ready to connect, set these properties:
- AuthScheme:
AccessKey. - AccessKey: The access key value you just retrieved from the Azure Portal.
Shared Access Signature (SAS)
To connect using a Shared Access Signature, you must first generate one using the Azure Storage Explorer tool.
When you are ready to connect, set these properties:
- AuthScheme:
SAS. - SharedAccessSignature: The value of the Shared Access Signature you just generated.
Entra ID (Microsoft Entra ID)
Note
Microsoft has rebranded Microsoft Entra ID as Entra ID. In topics that require the user to interact with the Entra ID Admin site, we use the same names Microsoft does. However, there are still connection properties whose names or values reference Microsoft Entra ID.
Microsoft Entra ID is a multi-tenant, cloud-based identity and access management platform. It supports OAuth-based authentication flows that enable the driver to access Azure Data Lake Storage endpoints securely.
Authentication to Entra ID via a web application always requires that you first create and register a custom OAuth application. This enables your application to define its own redirect URI, manage credential scope, and comply with organization-specific security policies.
For full instructions on how to create and register a custom OAuth application, see Creating an Entra ID (Microsoft Entra ID) Application.
After setting AuthScheme to Microsoft Entra ID, the steps to authenticate vary, depending on the environment. For details on how to connect from desktop applications, web-based workflows, or headless systems, see the following sections..
Desktop Applications
You can authenticate from a desktop application using either the driver's embedded OAuth application or a custom OAuth application registered in Microsoft Entra ID.
Option 1: Use the Embedded OAuth Application
This is a pre-registered application, included with the driver. It simplifies setup and eliminates the need to register your own credentials and is ideal for development environments, single-user tools, or any setup where quick and easy authentication is preferred.
Set the following connection properties:
- AuthScheme:
Microsoft Entra ID - InitiateOAuth:
GETANDREFRESH– Use for the initial login. Launches the login page and saves tokens.REFRESH– Use this setting when you have already obtained valid access and refresh tokens. Reuses stored tokens without prompting the user again.
When you connect, the driver opens the Microsoft Entra sign-in page in your default browser. After signing in and granting access, the driver retrieves the access and refresh tokens and saves them to the path specified by OAuthSettingsLocation.
Option 2: Use a Custom OAuth Application
If your organization requires more control, such as managing security policies, redirect URIs, or application branding, you can instead register a custom OAuth application in Microsoft Entra ID and provide its values during connection.
During registration, record the following values:
- OAuthClientId: The client ID that was generated when you registered your custom OAuth application.
- OAuthClientSecret: The client secret that was that was generated when you registered your custom OAuth application.
- CallbackURL: A redirect URI you defined during application registration.
For full instructions on how to register a custom OAuth application and configure redirect URIs, see Creating an Entra ID (Microsoft Entra ID) Application.
Set the following connection properties:
- AuthScheme:
Microsoft Entra ID - InitiateOAuth:
GETANDREFRESH– Use for the initial login. Launches the login page and saves tokens.REFRESH– Use this setting when you have already obtained valid access and refresh tokens. Reuses stored tokens without prompting the user again.
- OAuthClientId: The client ID that was generated when you registered your custom OAuth application.
- OAuthClientSecret: The client secret that was generated when you registered your custom OAuth application.
- CallbackURL: A redirect URI you defined during application registration.
After authentication, tokens are saved to OAuthSettingsLocation. These values persist across sessions and are used to automatically refresh the access token when it expires, so you don't need to log in again on future connections.
Microsoft Entra service principal
Note
Microsoft has rebranded Microsoft Entra ID as Entra ID. In topics that require the user to interact with the Entra ID Admin site, we use the same names Microsoft does. However, there are still connection properties whose names or values reference Microsoft Entra ID.
Service principals are security objects within a Microsoft Entra ID (Microsoft Entra ID) application that define what that application can do within a specific tenant. Service principals are created in the Entra admin center, also accessible through the Azure portal. As part of the creation process we also specify whether the service principal will access Entra resources via a client secret or a certificate.
Instead of being tied to a particular user, service principal permissions are based on the roles assigned to them. These roles determine which resources the application can access and which operations it can perform.
When authenticating using a service principal, you must register an application with an Entra tenant, as described in Creating a Service Principal App in Entra ID (Microsoft Entra ID).
This subsection describes properties you must set before you can connect. These vary, depending on whether you will authenticate via a client secret or a certificate.
Authentication with Client Secret
- AuthScheme:
AzureServicePrincipal. - AzureTenant: The Microsoft Entra ID tenant to which you will connect.
- OAuthClientId: The client ID in your application settings.
- OAuthClientSecret: The client secret in your application settings.
- InitiateOAuth:
GETANDREFRESH. You can use InitiateOAuth to avoid repeating the OAuth exchange and manually setting the OAuthAccessToken.
Authentication with Certificate
- AuthScheme:
AzureServicePrincipalCert. - AzureTenant: The Microsoft Entra ID tenant to which you will connect.
- OAuthClientId: The client ID in your application settings.
- OAuthJWTCert: The JWT Certificate store.
- OAuthJWTCertType: The JWT Certificate store type.
- InitiateOAuth:
GETANDREFRESH. You can use InitiateOAuth to avoid repeating the OAuth exchange and manually setting the OAuthAccessToken.
Managed Service Identity (MSI)
If you are running Azure Data Lake Storage on an Azure VM and want to automatically obtain Managed Service Identity (MSI) credentials to connect, set AuthScheme to AzureMSI.
User-Managed Identities
To obtain a token for a managed identity, use the OAuthClientId property to specify the managed identity's client_id.
If your VM has multiple user-assigned managed identities, you must also specify OAuthClientId.
Create an Entra ID (Microsoft Entra ID) Application
Create an Entra ID (Microsoft Entra ID) Application
Note
Microsoft has rebranded Microsoft Entra ID as Entra ID. In topics that require the user to interact with the Entra ID Admin site, we use the same names Microsoft does. However, there are still connection properties whose names or values reference Microsoft Entra ID.
Azure Data Lake Storage supports OAuth-based authentication using Microsoft Entra ID. If you will connect via a web application and want to authenticate via Entra ID, you must first register a custom OAuth application in the Entra Admin Center, as described below.
If you will connect via a desktop application or headless machine, you can authenticate using Azure Data Lake Storage's built-in embedded application credentials, which use branding. However, custom OAuth applications are also compatible with desktop and headless authentication flows, and may be preferable for production deployments or environments requiring strict policy control.
Register the Application
To register an OAuth application in Microsoft Entra ID, follow these steps:
-
Go to
https://portal.azure.com. -
In the left-hand navigation pane, select
Microsoft Entra ID > App registrations. -
Click
New registration. -
Enter a name for the application.
-
Specify the types of accounts this application should support:
- For private-use applications, select
Accounts in this organization directory only. - For distributed applications, select one of the multi-tenant options.
- For private-use applications, select
Note
If you select Accounts in this organizational directory only, when you connect with Azure Data Lake Storage connector, you must set AzureTenant to the tenant's ID (either GUID or verified domain). Otherwise, authentication will fail.
-
Set the redirect URI to
http://localhost:33333(default), or use another URI appropriate for your deployment.When using a custom redirect URI set a
CallbackURLconnection property; in those cases, set it to match this URI exactly. -
Click
Register. The application management screen opens. Record these values for later use:Application (client) IDis used for OAuthClientIdDirectory (tenant) IDis used for AzureTenant
-
Go to
Certificates & Secrets. ClickNew Client Secret, set the desired expiration, and save the generated value. This value will only be shown once — record it to use with OAuthClientSecret. -
Select 'Windows Azure Service Management API' along with 'Azure Data Lake' and add the
Have full access to the Azure Data Lake servicepermission. -
To confirm, click
Add permissions.
Create a Service Principal App in Entra ID (Microsoft Entra ID)
Create a Service Principal App in Entra ID (Microsoft Entra ID)
Note
Microsoft has rebranded Microsoft Entra ID as Entra ID. In topics that require the user to interact with the Entra ID Admin site, we use the same names Microsoft does. However, there are still connection properties whose names or values reference Microsoft Entra ID.
Azure Data Lake Storage supports Service Principal-based authentication, which is role-based. This means that the Service Principal's permissions are determined by the roles assigned to it. The roles specify what resources the Service Principal can access and which operations it can perform.
If you want to use a Service Principal to authenticate to Azure Data Lake Storage, you must create a custom application in Microsoft Entra ID.
To enable Service Principal authentication:
- Confirm that you have permission to register applications and assign roles in your tenant.
- Register a new application and configure credentials and permissions in the Entra Admin Center.
Register the Application
- Go to
https://portal.azure.com. - In the left-hand navigation pane, select
Microsoft Entra ID > App registrations. - Click
New registration. - Enter a name for the application.
- Select the desired tenant setup. Since this custom application is for Service Principal use, choose
Any Microsoft Entra ID tenant – Multitenant. - Click
Register. The application management screen opens. Note the value inApplication (client) IDas the OAuthClientId and theDirectory (tenant) IDas the AzureTenant - Navigate to
Certificates & Secretsand define the application authentication type. Two types of authentication are available: certificate (recommended) or client secret- For certificate authentication: In
Certificates & Secrets, selectUpload certificate, then upload the certificate from your local machine. For more information on creating a self-signed certificate, see Create a self-signed certificate - For creating a new client secret: In
Certificates & Secrets, selectNew Client Secretfor the application and specify its duration. After the client secret is saved, Azure Data Lake Storage displays the key value. This value is displayed only once, so be sure to record it for future use. Use this value for the OAuthClientSecret
- For certificate authentication: In
- Navigate to
Authenticationand select theAccess tokensoption. - Select 'Windows Azure Service Management API' along with 'Azure Data Lake' and add the
Have full access to the Azure Data Lake servicepermission. - Save your changes.
- If you specified permissions that require admin consent (such as the Application Permissions), you can grant them from the current tenant on the API Permissions page.
Grant Admin Consent
Some custom applications require administrative permissions to operate within a Microsoft Entra ID tenant. This is especially true for applications that use Application permissions, which allow the app to run without a signed-in user. Admin consent can be granted when creating a new application, by adding relevant permissions marked as "Admin Consent Required".
Admin consent is also required to use Client Credentials in the authentication flow.
These permissions must be granted by an admin. To grant admin consent:
- Log in to
https://portal.azure.comwith an administrator account. - Navigate to
Microsoft Entra ID > App registrationsand select your registered application. - Navigate to
API permissions. - Review the permissions listed under
Application permissions. Ensure the necessary API scopes are included for your use case. - Click
Grant admin consentto approve the requested permissions.
This gives your application permissions on the tenant under which it was created.
Consent for Client Credentials
OAuth supports the use of client credentials to authenticate. In a client credentials authentication flow, credentials are created for the authenticating application itself. The auth flow acts just like the usual auth flow, except that there is no prompt for an associated user to provide credentials. All tasks accepted by the application are executed outside of the context of a default user.
Note
Since the embedded OAuth credentials authenticate on a per-user basis, you cannot use them in a client authentication flow. You must always create a custom OAuth application to use client credentials.
- Log in to
https://portal.azure.com - Create a custom OAuth application, as described above.
- Navigate to
App Registrations. - Find the application you just created, and open
API Permissions. - Select the Microsoft Graph permissions. There are two distinct sets of permissions: Delegated and Application.
- For use with Service Principal, specify
Applicationpermissions. - Select the permissions you require for your integration.
Fine-Tuning Data Access
Fine Tuning Data Access
To gain more control over the data returned from Azure Data Lake Storage Gen2's ADLSGen2 schema, set IncludeSubDirectories. This property specifies whether the connector should list the subdirectories' paths in the Resources view.
Important Notes
Configuration Files and Their Paths
- All references to adding configuration files and their paths refer to files and locations on the Jitterbit agent where the connector is installed. These paths are to be adjusted as appropriate depending on the agent and the operating system. If multiple agents are used in an agent group, identical files will be required on each agent.
Advanced Features
This section details a selection of advanced features of the Azure Data Lake Storage connector.
User Defined Views
The connector supports the use of user defined views, virtual tables whose contents are decided by a pre-configured user defined query. These views are useful when you cannot directly control queries being issued to the drivers. For an overview of creating and configuring custom views, see User Defined Views.
SSL Configuration
Use SSL Configuration to adjust how connector handles TLS/SSL certificate negotiations. You can choose from various certificate formats. For further information, see the SSLServerCert property under "Connection String Options".
Proxy
To configure the connector using private agent proxy settings, select the Use Proxy Settings checkbox on the connection configuration screen.
Query Processing
The connector offloads as much of the SELECT statement processing as possible to Azure Data Lake Storage and then processes the rest of the query in memory (client-side).
For further information, see Query Processing.
Log
For an overview of configuration settings that can be used to refine logging, see Logging. Only two connection properties are required for basic logging, but there are numerous features that support more refined logging, which enables you to use the LogModules connection property to specify subsets of information to be logged.
User Defined Views
The Azure Data Lake Storage connector supports the use of user defined views: user-defined virtual tables whose contents are decided by a preconfigured query. User defined views are useful in situations where you cannot directly control the query being issued to the driver; for example, when using the driver from Jitterbit.
Use a user defined view to define predicates that are always applied. If you specify additional predicates in the query to the view, they are combined with the query already defined as part of the view.
There are two ways to create user defined views:
- Create a JSON-formatted configuration file defining the views you want.
- DDL statements.
Define Views Using a Configuration File
User defined views are defined in a JSON-formatted configuration file called UserDefinedViews.json. The connector automatically detects the views specified in this file.
You can also have multiple view definitions and control them using the UserDefinedViews connection property. When you use this property, only the specified views are seen by the connector.
This user defined view configuration file is formatted so that each root element defines the name of a view, and includes a child element, called query, which contains the custom SQL query for the view.
For example:
{
"MyView": {
"query": "SELECT * FROM Resources WHERE MyColumn = 'value'"
},
"MyView2": {
"query": "SELECT * FROM MyTable WHERE Id IN (1,2,3)"
}
}
Use the UserDefinedViews connection property to specify the location of your JSON configuration file. For example:
"UserDefinedViews", "C:\Users\yourusername\Desktop\tmp\UserDefinedViews.json"
Define Views Using DDL Statements
The connector is also capable of creating and altering the schema via DDL Statements such as CREATE LOCAL VIEW, ALTER LOCAL VIEW, and DROP LOCAL VIEW.
Create a View
To create a new view using DDL statements, provide the view name and query as follows:
CREATE LOCAL VIEW [MyViewName] AS SELECT * FROM Customers LIMIT 20;
If no JSON file exists, the above code creates one. The view is then created in the JSON configuration file and is now discoverable. The JSON file location is specified by the UserDefinedViews connection property.
Alter a View
To alter an existing view, provide the name of an existing view alongside the new query you would like to use instead:
ALTER LOCAL VIEW [MyViewName] AS SELECT * FROM Customers WHERE TimeModified > '3/1/2020';
The view is then updated in the JSON configuration file.
Drop a View
To drop an existing view, provide the name of an existing schema alongside the new query you would like to use instead.
DROP LOCAL VIEW [MyViewName]
This removes the view from the JSON configuration file. It can no longer be queried.
Schema for User Defined Views
In order to avoid a view's name clashing with an actual entity in the data model, user defined views are exposed in the UserViews schema by default. To change the name of the schema used for UserViews, reset the UserViewsSchemaName property.
Work with User Defined Views
For example, a SQL statement with a user defined view called UserViews.RCustomers only lists customers in Raleigh:
SELECT * FROM Customers WHERE City = 'Raleigh';
An example of a query to the driver:
SELECT * FROM UserViews.RCustomers WHERE Status = 'Active';
Resulting in the effective query to the source:
SELECT * FROM Customers WHERE City = 'Raleigh' AND Status = 'Active';
That is a very simple example of a query to a user defined view that is effectively a combination of the view query and the view definition. It is possible to compose these queries in much more complex patterns. All SQL operations are allowed in both queries and are combined when appropriate.
SSL Configuration
Customize the SSL Configuration
By default, the connector attempts to negotiate TLS with the server. The server certificate is validated against the default system trusted certificate store. You can override how the certificate gets validated using the SSLServerCert connection property.
To specify another certificate, see the SSLServerCert connection property.
Data Model
Overview
This section shows the available API objects and provides more information on executing SQL to Azure Data Lake Storage APIs.
Key Features
- The connector models Azure Data Lake Storage entities like documents, folders, and groups as relational views, allowing you to write SQL to query Azure Data Lake Storage data.
- Stored procedures allow you to execute operations to Azure Data Lake Storage.
- Live connectivity to these objects means any changes to your Azure Data Lake Storage account are immediately reflected when using the connector.
Views
Views describes the available views. Views are statically defined to model Resources and Permissions.
Stored Procedures
Stored Procedures are function-like interfaces to Azure Data Lake Storage. Stored procedures allow you to execute operations to Azure Data Lake Storage, including downloading documents and moving envelopes.
Views
Views are similar to tables in the way that data is represented; however, views are read-only.
Queries can be executed against a view as if it were a normal table.
Azure Data Lake Storage Connector Views
| Name | Description |
|---|---|
FileSystems |
Lists all filesystems. |
Resources |
Lists the contents of the supplied path. |
FileSystems
Lists all filesystems.
Columns
| Name | Type | Behavior | Description |
|---|---|---|---|
ETag |
String |
Unique identifier of the filesystem. | |
Name [KEY] |
String |
The name of the filesystem. | |
LastModified |
Timestamp |
The timestamp when the filesystem was modified for the last time. |
Resources
Lists the contents of the supplied path.
Columns
| Name | Type | Behavior | Description |
|---|---|---|---|
Name [KEY] |
String |
The path of the file or folder. | |
Directory |
String |
The directory path of the file or folder. | |
IsDirectory |
String |
Determines if the resource is a folder or a file. | |
ContentLength |
Long |
Determines the size of the file in bytes. | |
LastModified |
Timestamp |
The timestamp when the file was modified for the last time. | |
Owner |
String |
The name of the owner. | |
Permissions |
String |
The permissions set to the file. | |
ETag |
String |
Unique identifier of the file or folder. |
Pseudo-Columns
Pseudo column fields are used in the WHERE clause of SELECT statements and offer a more granular control over the tuples that are returned from the data source.
| Name | Type | Description |
|---|---|---|
Recursive |
Boolean |
Set this to true to retrieve all sub folders and files. The default value is false. |
Stored Procedures
Stored procedures are function-like interfaces that extend the functionality of the connector beyond simple SELECT operations with Azure Data Lake Storage.
Stored procedures accept a list of parameters, perform their intended function, and then return any relevant response data from Azure Data Lake Storage, along with an indication of whether the procedure succeeded or failed.
Azure Data Lake Storage Connector Stored Procedures
| Name | Description |
|---|---|
CopyFile |
Copy a file from a source to a destination. |
CreateFile |
Create a file. Path is required for a successful operation. |
CreateFileSystem |
Create a filesystem by specifying the name. |
CreateFolder |
Create a folder. Path is required for a successful operation. |
DeleteFileSystem |
Delete a filesystem. The name of the filesystem for a successful operation. |
DeleteObject |
Delete a file or a folder. Path is required for a successful operation. |
DownloadFile |
Download a file. Path and DownloadPath are required for a successful operation. |
GetFileSystemProperties |
Get the properties of a filesystem. The name of the filesystem is required for a successful operation. |
GetOAuthAccessToken |
Gets the OAuth access token from SharePoint. |
GetOAuthAuthorizationURL |
Gets the SharePoint authorization URL. Access the URL returned in the output in a Web browser. This requests the access token that can be used as part of the connection string to SharePoint. |
LeaseBlob |
The LeaseBlob operation creates and manages a lock on a blob for write and delete operations. |
RefreshOAuthAccessToken |
Refreshes the OAuth access token used for authentication with SharePoint. |
RenameObject |
Rename a file or a directory. Path and RenameTo are required for a successful operation. |
SetFileSystemProperties |
Set the properties for a filesystem. The name of the filesystem is required for a successful operation. |
UploadFile |
Upload a File. Path and FilePath are required for a successful operation. |
CopyFile
Copy a file from a source to a destination.
Input
| Name | Type | Description |
|---|---|---|
SourcePath |
String |
Required. The path of the file which will be copied. |
DestinationPath |
String |
Required. The path of the file where it will be copied. |
Result Set Columns
| Name | Type | Description |
|---|---|---|
Success |
String |
Determines if the operation was successful or not. |
CreateFile
Create a file. Path is required for a successful operation.
Input
| Name | Type | Description |
|---|---|---|
Path |
String |
The path of the file which will be created. |
Result Set Columns
| Name | Type | Description |
|---|---|---|
Success |
String |
Determines if the operation was successful or not. |
CreateFileSystem
Create a filesystem by specifying the name.
Input
| Name | Type | Description |
|---|---|---|
FileSystem |
String |
The name for the new filesystem. |
Properties |
String |
Properties to set to the filesystem. |
DefaultEncryptionScope |
String |
Default encryption scopes to set to the filesystem. |
Result Set Columns
| Name | Type | Description |
|---|---|---|
Success |
String |
Determines if the operation was successful or not. |
ETag |
String |
Unique identifier of the filesystem. |
CreateFolder
Create a folder. Path is required for a successful operation.
Input
| Name | Type | Description |
|---|---|---|
Path |
String |
The path of the folder which will be created. |
Result Set Columns
| Name | Type | Description |
|---|---|---|
Success |
String |
Determines if the operation was successful or not. |
DeleteFileSystem
Delete a filesystem. The name of the filesystem for a successful operation.
Input
| Name | Type | Description |
|---|---|---|
FileSystem |
String |
The name of the filesystem which will be deleted. |
ModifiedSince |
Datetime |
A date and time value that sets condition to perform the operation only if the resource has been modified since the specified date and time. |
UnmodifiedSince |
Datetime |
A date and time value that sets condition to perform the operation only if the resource has not been modified since the specified date and time. |
Result Set Columns
| Name | Type | Description |
|---|---|---|
Success |
Boolean |
Determines if the operation was successful or not. |
DeleteObject
Delete a file or a folder. Path is required for a successful operation.
Input
| Name | Type | Description |
|---|---|---|
Path |
String |
The path of the file or folder to be deleted. |
Recursive |
Boolean |
Set this to true to delete a folder with all of it content. The default value is false. |
DeleteType |
String |
Set this to FILESANDFOLDERS to delete the file or folder specified in the Path parameter. Set this to FILES to delete only the files inside the folder specified in the Path parameter. The allowed values are FILES, FILESANDFOLDERS. The default value is FILESANDFOLDERS. |
Result Set Columns
| Name | Type | Description |
|---|---|---|
Success |
String |
Determines if the operation was successful or not. |
DownloadFile
Download a file. Path and DownloadPath are required for a successful operation.
Input
| Name | Type | Description |
|---|---|---|
Path |
String |
The path of the file which will be downloaded. |
DownloadPath |
String |
The path where the file will be downloaded. |
Result Set Columns
| Name | Type | Description |
|---|---|---|
FileData |
String |
File data that will be outputted encoded in Base64 if the DownloadPath and FileStream inputs are empty. |
Success |
String |
Whether the operation was successful. |
GetFileSystemProperties
Get the properties of a filesystem. The name of the filesystem is required for a successful operation.
Input
| Name | Type | Description |
|---|---|---|
FileSystem |
String |
The path of the file which will be created. |
Result Set Columns
| Name | Type | Description |
|---|---|---|
Success |
String |
Determines if the operation was successful or not. |
ETag |
String |
Unique identifier of the filesystem. |
Properties |
String |
Properties of the filesystem. |
DefaultEncryptionScope |
String |
Default encryption scopes of the filesystem. |
DenyEncryptionScopeOverride |
Boolean |
Determines if encryption scope override is denied for the filesystem. |
LastModified |
Datetime |
The timestamp when the filesystem was modified for the last time. |
GetOAuthAccessToken
Gets the OAuth access token from SharePoint.
Input
| Name | Type | Description |
|---|---|---|
AuthMode |
String |
The type of authentication mode to use. The allowed values are APP, WEB. |
Verifier |
String |
The verifier token returned by SharePoint after using the URL obtained with GetOAuthAuthorizationURL. Required for only the Web AuthMode. |
CallbackUrl |
String |
The URL the user will be redirected to after authorizing your application. |
State |
String |
Any value that you wish to be sent with the callback. |
Prompt |
String |
Defaults to 'select_account' which prompts the user to select account while authenticating. Set to 'None', for no prompt, 'login' to force user to enter their credentials or 'consent' to trigger the OAuth consent dialog after the user signs in, asking the user to grant permissions to the app. |
Result Set Columns
| Name | Type | Description |
|---|---|---|
OAuthAccessToken |
String |
The authentication token returned from SharePoint. |
OAuthRefreshToken |
String |
A token that may be used to obtain a new access token. |
ExpiresIn |
String |
The remaining lifetime for the access token in seconds. |
GetOAuthAuthorizationURL
Gets the SharePoint authorization URL. Access the URL returned in the output in a Web browser. This requests the access token that can be used as part of the connection string to SharePoint.
Input
| Name | Type | Description |
|---|---|---|
CallbackUrl |
String |
The URL that Sharepoint will return to after the user has authorized your app. |
State |
String |
Any value that you wish to be sent with the callback. |
Prompt |
String |
Defaults to 'select_account' which prompts the user to select account while authenticating. Set to 'None', for no prompt, 'login' to force user to enter their credentials or 'consent' to trigger the OAuth consent dialog after the user signs in, asking the user to grant permissions to the app. |
Result Set Columns
| Name | Type | Description |
|---|---|---|
URL |
String |
The URL to be entered into a Web browser to obtain the verifier token and authorize the data provider with. |
LeaseBlob
The LeaseBlob operation creates and manages a lock on a blob for write and delete operations.
Input
| Name | Type | Description |
|---|---|---|
Path |
String |
The path of the file which will be created. |
LeaseAction |
String |
Specifies the lease action to execute The allowed values are Acquire, Renew, Change, Release, Break. |
LeaseId |
String |
Required to Renew, Change, or Release the lease. |
LeaseBreak |
Integer |
For a Break operation, this is the proposed duration of seconds that the lease should continue before it is broken, between 0 and 60 seconds. |
LeaseDuration |
Integer |
Only allowed and required on an Acquire operation. Specifies the duration of the lease, in seconds, or negative one (-1) for a lease that never expires. A non-infinite lease can be between 15 and 60 seconds. A lease duration can't be changed by using renew or change. |
ProposedLeaseId |
String |
Optional for Acquire, and required for Change. Proposed lease ID, in a GUID string format. |
Result Set Columns
| Name | Type | Description |
|---|---|---|
LeaseId |
String |
The ID of the lease. |
Success |
String |
Determines if the operation was successful or not. |
RefreshOAuthAccessToken
Refreshes the OAuth access token used for authentication with SharePoint.
Input
| Name | Type | Description |
|---|---|---|
OAuthRefreshToken |
String |
The old token to be refreshed. |
Result Set Columns
| Name | Type | Description |
|---|---|---|
OAuthAccessToken |
String |
The authentication token returned from SharePoint. |
ExpiresIn |
String |
The remaining lifetime on the access token. |
RenameObject
Rename a file or a directory. Path and RenameTo are required for a successful operation.
Input
| Name | Type | Description |
|---|---|---|
Path |
String |
The path which will be renamed. |
RenameTo |
String |
The new name of the file/folder. |
Result Set Columns
| Name | Type | Description |
|---|---|---|
Success |
Boolean |
Whether the operation completed successfully or not. |
SetFileSystemProperties
Set the properties for a filesystem. The name of the filesystem is required for a successful operation.
Input
| Name | Type | Description |
|---|---|---|
FileSystem |
String |
The name of the filesystem for which the properties will be set. |
Properties |
String |
User-defined properties to be stored with the filesystem, in the format of a comma-separated list of name and value pairs 'n1=v1, n2=v2, ...', where each value is a base64 encoded string. Note that the string may only contain ASCII characters in the ISO-8859-1 character set. If the filesystem exists, any properties not included in the list will be removed. |
ETag |
String |
Unique identifier of the filesystem used to set a condition to perform the operation on the basis that it matches with the current value of the filesystem. Only one conditional parameter is allowed per operation. |
ModifiedSince |
Datetime |
A date and time value that sets condition to perform the operation only if the resource has been modified since the specified date and time. Only one conditional parameter is allowed per operation. |
UnmodifiedSince |
Datetime |
A date and time value that sets condition to perform the operation only if the resource has not been modified since the specified date and time. Only one conditional parameter is allowed per operation. |
Result Set Columns
| Name | Type | Description |
|---|---|---|
Success |
String |
Determines if the operation was successful or not. |
ETag |
String |
Unique identifier of the filesystem. |
UploadFile
Upload a File. Path and FilePath are required for a successful operation.
Stored Procedure Specific Information
The behavior of this stored procedure depends on the nature of the FilePath parameter and is influenced by the MaxThreads connection property, which defines the maximum number of threads to use during the upload process:
- If
FilePathpoints to asingle file, the driver uploads it inparallel chunks. The size of each chunk is determined by the ChunkSize connection property, and the upload is distributed across multiple threads based on the value of MaxThreads. - If
FilePathpoints to adirectory, the driver uploads each file in the directoryin parallel, usingone thread per file. In this mode, multipart chunking is not applied to individual files.
Examples:
-
Uploading a single file in parallel using multiple threads and chunking:
EXECUTE UploadFile FilePath = 'C:/Users/User/Desktop/SampleUploadtest.txt' -
Uploading all files in a directory in parallel, one thread per file (no chunking):
EXECUTE UploadFile FilePath = 'C:/Users/User/Desktop/'
Input
| Name | Type | Description |
|---|---|---|
Path |
String |
Set this to the path of the file where it will be uploaded. |
FilePath |
String |
Set this to the path of the file which will be uploaded. Example: C:/Users/User/Desktop/SampleUploadtest.txt. This parameter takes priority over others. |
FileContent |
String |
Base64 encoded content of the file to be added. |
Result Set Columns
| Name | Type | Description |
|---|---|---|
Success |
Boolean |
Whether the operation completed successfully or not. |
System Tables
You can query the system tables described in this section to access schema information, information on data source functionality, and batch operation statistics.
Schema Tables
The following tables return database metadata for Azure Data Lake Storage:
- sys_catalogs: Lists the available databases.
- sys_schemas: Lists the available schemas.
- sys_tables: Lists the available tables and views.
- sys_tablecolumns: Describes the columns of the available tables and views.
- sys_procedures: Describes the available stored procedures.
- sys_procedureparameters: Describes stored procedure parameters.
- sys_keycolumns: Describes the primary and foreign keys.
- sys_indexes: Describes the available indexes.
Data Source Tables
The following tables return information about how to connect to and query the data source:
- sys_connection_props: Returns information on the available connection properties.
- sys_sqlinfo: Describes the SELECT queries that the connector can offload to the data source.
Query Information Tables
The following table returns query statistics for data modification queries:
- sys_identity: Returns information about batch operations or single updates.
sys_catalogs
Lists the available databases.
The following query retrieves all databases determined by the connection string:
SELECT * FROM sys_catalogs
Columns
| Name | Type | Description |
|---|---|---|
CatalogName |
String |
The database name. |
sys_schemas
Lists the available schemas.
The following query retrieves all available schemas:
SELECT * FROM sys_schemas
Columns
| Name | Type | Description |
|---|---|---|
CatalogName |
String |
The database name. |
SchemaName |
String |
The schema name. |
sys_tables
Lists the available tables.
The following query retrieves the available tables and views:
SELECT * FROM sys_tables
Columns
| Name | Type | Description |
|---|---|---|
CatalogName |
String |
The database containing the table or view. |
SchemaName |
String |
The schema containing the table or view. |
TableName |
String |
The name of the table or view. |
TableType |
String |
The table type (table or view). |
Description |
String |
A description of the table or view. |
IsUpdateable |
Boolean |
Whether the table can be updated. |
sys_tablecolumns
Describes the columns of the available tables and views.
The following query returns the columns and data types for the Resources table:
SELECT ColumnName, DataTypeName FROM sys_tablecolumns WHERE TableName='Resources'
Columns
| Name | Type | Description |
|---|---|---|
CatalogName |
String |
The name of the database containing the table or view. |
SchemaName |
String |
The schema containing the table or view. |
TableName |
String |
The name of the table or view containing the column. |
ColumnName |
String |
The column name. |
DataTypeName |
String |
The data type name. |
DataType |
Int32 |
An integer indicating the data type. This value is determined at run time based on the environment. |
Length |
Int32 |
The storage size of the column. |
DisplaySize |
Int32 |
The designated column's normal maximum width in characters. |
NumericPrecision |
Int32 |
The maximum number of digits in numeric data. The column length in characters for character and date-time data. |
NumericScale |
Int32 |
The column scale or number of digits to the right of the decimal point. |
IsNullable |
Boolean |
Whether the column can contain null. |
Description |
String |
A brief description of the column. |
Ordinal |
Int32 |
The sequence number of the column. |
IsAutoIncrement |
String |
Whether the column value is assigned in fixed increments. |
IsGeneratedColumn |
String |
Whether the column is generated. |
IsHidden |
Boolean |
Whether the column is hidden. |
IsArray |
Boolean |
Whether the column is an array. |
IsReadOnly |
Boolean |
Whether the column is read-only. |
IsKey |
Boolean |
Indicates whether a field returned from sys_tablecolumns is the primary key of the table. |
ColumnType |
String |
The role or classification of the column in the schema. Possible values include SYSTEM, LINKEDCOLUMN, NAVIGATIONKEY, REFERENCECOLUMN, and NAVIGATIONPARENTCOLUMN. |
sys_procedures
Lists the available stored procedures.
The following query retrieves the available stored procedures:
SELECT * FROM sys_procedures
Columns
| Name | Type | Description |
|---|---|---|
CatalogName |
String |
The database containing the stored procedure. |
SchemaName |
String |
The schema containing the stored procedure. |
ProcedureName |
String |
The name of the stored procedure. |
Description |
String |
A description of the stored procedure. |
ProcedureType |
String |
The type of the procedure, such as PROCEDURE or FUNCTION. |
sys_procedureparameters
Describes stored procedure parameters.
The following query returns information about all of the input parameters for the DownloadFile stored procedure:
SELECT * FROM sys_procedureparameters WHERE ProcedureName = 'DownloadFile' AND Direction = 1 OR Direction = 2
To include result set columns in addition to the parameters, set the IncludeResultColumns pseudo column to True:
SELECT * FROM sys_procedureparameters WHERE ProcedureName = 'DownloadFile' AND IncludeResultColumns='True'
Columns
| Name | Type | Description |
|---|---|---|
CatalogName |
String |
The name of the database containing the stored procedure. |
SchemaName |
String |
The name of the schema containing the stored procedure. |
ProcedureName |
String |
The name of the stored procedure containing the parameter. |
ColumnName |
String |
The name of the stored procedure parameter. |
Direction |
Int32 |
An integer corresponding to the type of the parameter: input (1), input/output (2), or output(4). input/output type parameters can be both input and output parameters. |
DataType |
Int32 |
An integer indicating the data type. This value is determined at run time based on the environment. |
DataTypeName |
String |
The name of the data type. |
NumericPrecision |
Int32 |
The maximum precision for numeric data. The column length in characters for character and date-time data. |
Length |
Int32 |
The number of characters allowed for character data. The number of digits allowed for numeric data. |
NumericScale |
Int32 |
The number of digits to the right of the decimal point in numeric data. |
IsNullable |
Boolean |
Whether the parameter can contain null. |
IsRequired |
Boolean |
Whether the parameter is required for execution of the procedure. |
IsArray |
Boolean |
Whether the parameter is an array. |
Description |
String |
The description of the parameter. |
Ordinal |
Int32 |
The index of the parameter. |
Values |
String |
The values you can set in this parameter are limited to those shown in this column. Possible values are comma-separated. |
SupportsStreams |
Boolean |
Whether the parameter represents a file that you can pass as either a file path or a stream. |
IsPath |
Boolean |
Whether the parameter is a target path for a schema creation operation. |
Default |
String |
The value used for this parameter when no value is specified. |
SpecificName |
String |
A label that, when multiple stored procedures have the same name, uniquely identifies each identically-named stored procedure. If there's only one procedure with a given name, its name is simply reflected here. |
IsProvided |
Boolean |
Whether the procedure is added/implemented by , as opposed to being a native Azure Data Lake Storage procedure. |
Pseudo-Columns
| Name | Type | Description |
|---|---|---|
IncludeResultColumns |
Boolean |
Whether the output should include columns from the result set in addition to parameters. Defaults to False. |
sys_keycolumns
Describes the primary and foreign keys.
The following query retrieves the primary key for the Resources table:
SELECT * FROM sys_keycolumns WHERE IsKey='True' AND TableName='Resources'
Columns
| Name | Type | Description |
|---|---|---|
CatalogName |
String |
The name of the database containing the key. |
SchemaName |
String |
The name of the schema containing the key. |
TableName |
String |
The name of the table containing the key. |
ColumnName |
String |
The name of the key column. |
IsKey |
Boolean |
Whether the column is a primary key in the table referenced in the TableName field. |
IsForeignKey |
Boolean |
Whether the column is a foreign key referenced in the TableName field. |
PrimaryKeyName |
String |
The name of the primary key. |
ForeignKeyName |
String |
The name of the foreign key. |
ReferencedCatalogName |
String |
The database containing the primary key. |
ReferencedSchemaName |
String |
The schema containing the primary key. |
ReferencedTableName |
String |
The table containing the primary key. |
ReferencedColumnName |
String |
The column name of the primary key. |
sys_foreignkeys
Describes the foreign keys.
The following query retrieves all foreign keys which refer to other tables:
SELECT * FROM sys_foreignkeys WHERE ForeignKeyType = 'FOREIGNKEY_TYPE_IMPORT'
Columns
| Name | Type | Description |
|---|---|---|
CatalogName |
String |
The name of the database containing the key. |
SchemaName |
String |
The name of the schema containing the key. |
TableName |
String |
The name of the table containing the key. |
ColumnName |
String |
The name of the key column. |
PrimaryKeyName |
String |
The name of the primary key. |
ForeignKeyName |
String |
The name of the foreign key. |
ReferencedCatalogName |
String |
The database containing the primary key. |
ReferencedSchemaName |
String |
The schema containing the primary key. |
ReferencedTableName |
String |
The table containing the primary key. |
ReferencedColumnName |
String |
The column name of the primary key. |
ForeignKeyType |
String |
Designates whether the foreign key is an import (points to other tables) or export (referenced from other tables) key. |
sys_primarykeys
Describes the primary keys.
The following query retrieves the primary keys from all tables and views:
SELECT * FROM sys_primarykeys
Columns
| Name | Type | Description |
|---|---|---|
CatalogName |
String |
The name of the database containing the key. |
SchemaName |
String |
The name of the schema containing the key. |
TableName |
String |
The name of the table containing the key. |
ColumnName |
String |
The name of the key column. |
KeySeq |
String |
The sequence number of the primary key. |
KeyName |
String |
The name of the primary key. |
sys_indexes
Describes the available indexes. By filtering on indexes, you can write more selective queries with faster query response times.
The following query retrieves all indexes that are not primary keys:
SELECT * FROM sys_indexes WHERE IsPrimary='false'
Columns
| Name | Type | Description |
|---|---|---|
CatalogName |
String |
The name of the database containing the index. |
SchemaName |
String |
The name of the schema containing the index. |
TableName |
String |
The name of the table containing the index. |
IndexName |
String |
The index name. |
ColumnName |
String |
The name of the column associated with the index. |
IsUnique |
Boolean |
True if the index is unique. False otherwise. |
IsPrimary |
Boolean |
True if the index is a primary key. False otherwise. |
Type |
Int16 |
An integer value corresponding to the index type: statistic (0), clustered (1), hashed (2), or other (3). |
SortOrder |
String |
The sort order: A for ascending or D for descending. |
OrdinalPosition |
Int16 |
The sequence number of the column in the index. |
sys_connection_props
Returns information on the available connection properties and those set in the connection string.
The following query retrieves all connection properties that have been set in the connection string or set through a default value:
SELECT * FROM sys_connection_props WHERE Value <> ''
Columns
| Name | Type | Description |
|---|---|---|
Name |
String |
The name of the connection property. |
ShortDescription |
String |
A brief description. |
Type |
String |
The data type of the connection property. |
Default |
String |
The default value if one is not explicitly set. |
Values |
String |
A comma-separated list of possible values. A validation error is thrown if another value is specified. |
Value |
String |
The value you set or a preconfigured default. |
Required |
Boolean |
Whether the property is required to connect. |
Category |
String |
The category of the connection property. |
IsSessionProperty |
String |
Whether the property is a session property, used to save information about the current connection. |
Sensitivity |
String |
The sensitivity level of the property. This informs whether the property is obfuscated in logging and authentication forms. |
PropertyName |
String |
A camel-cased truncated form of the connection property name. |
Ordinal |
Int32 |
The index of the parameter. |
CatOrdinal |
Int32 |
The index of the parameter category. |
Hierarchy |
String |
Shows dependent properties associated that need to be set alongside this one. |
Visible |
Boolean |
Informs whether the property is visible in the connection UI. |
ETC |
String |
Various miscellaneous information about the property. |
sys_sqlinfo
Describes the SELECT query processing that the connector can offload to the data source.
Discovering the Data Source's SELECT Capabilities
Below is an example data set of SQL capabilities. Some aspects of SELECT functionality are returned in a comma-separated list if supported; otherwise, the column contains NO.
| Name | Description | Possible Values |
|---|---|---|
AGGREGATE_FUNCTIONS |
Supported aggregation functions. | AVG, COUNT, MAX, MIN, SUM, DISTINCT |
COUNT |
Whether COUNT function is supported. | YES, NO |
IDENTIFIER_QUOTE_OPEN_CHAR |
The opening character used to escape an identifier. | [ |
IDENTIFIER_QUOTE_CLOSE_CHAR |
The closing character used to escape an identifier. | ] |
SUPPORTED_OPERATORS |
A list of supported SQL operators. | =, >, <, >=, <=, <>, !=, LIKE, NOT LIKE, IN, NOT IN, IS NULL, IS NOT NULL, AND, OR |
GROUP_BY |
Whether GROUP BY is supported, and, if so, the degree of support. | NO, NO_RELATION, EQUALS_SELECT, SQL_GB_COLLATE |
STRING_FUNCTIONS |
Supported string functions. | LENGTH, CHAR, LOCATE, REPLACE, SUBSTRING, RTRIM, LTRIM, RIGHT, LEFT, UCASE, SPACE, SOUNDEX, LCASE, CONCAT, ASCII, REPEAT, OCTET, BIT, POSITION, INSERT, TRIM, UPPER, REGEXP, LOWER, DIFFERENCE, CHARACTER, SUBSTR, STR, REVERSE, PLAN, UUIDTOSTR, TRANSLATE, TRAILING, TO, STUFF, STRTOUUID, STRING, SPLIT, SORTKEY, SIMILAR, REPLICATE, PATINDEX, LPAD, LEN, LEADING, KEY, INSTR, INSERTSTR, HTML, GRAPHICAL, CONVERT, COLLATION, CHARINDEX, BYTE |
NUMERIC_FUNCTIONS |
Supported numeric functions. | ABS, ACOS, ASIN, ATAN, ATAN2, CEILING, COS, COT, EXP, FLOOR, LOG, MOD, SIGN, SIN, SQRT, TAN, PI, RAND, DEGREES, LOG10, POWER, RADIANS, ROUND, TRUNCATE |
TIMEDATE_FUNCTIONS |
Supported date/time functions. | NOW, CURDATE, DAYOFMONTH, DAYOFWEEK, DAYOFYEAR, MONTH, QUARTER, WEEK, YEAR, CURTIME, HOUR, MINUTE, SECOND, TIMESTAMPADD, TIMESTAMPDIFF, DAYNAME, MONTHNAME, CURRENT_DATE, CURRENT_TIME, CURRENT_TIMESTAMP, EXTRACT |
REPLICATION_SKIP_TABLES |
Indicates tables skipped during replication. | |
REPLICATION_TIMECHECK_COLUMNS |
A string array containing a list of columns which will be used to check for (in the given order) to use as a modified column during replication. | |
IDENTIFIER_PATTERN |
String value indicating what string is valid for an identifier. | |
SUPPORT_TRANSACTION |
Indicates if the provider supports transactions such as commit and rollback. | YES, NO |
DIALECT |
Indicates the SQL dialect to use. | |
KEY_PROPERTIES |
Indicates the properties which identify the uniform database. | |
SUPPORTS_MULTIPLE_SCHEMAS |
Indicates if multiple schemas may exist for the provider. | YES, NO |
SUPPORTS_MULTIPLE_CATALOGS |
Indicates if multiple catalogs may exist for the provider. | YES, NO |
DATASYNCVERSION |
The Data Sync version needed to access this driver. | Standard, Starter, Professional, Enterprise |
DATASYNCCATEGORY |
The Data Sync category of this driver. | Source, Destination, Cloud Destination |
SUPPORTSENHANCEDSQL |
Whether enhanced SQL functionality beyond what is offered by the API is supported. | TRUE, FALSE |
SUPPORTS_BATCH_OPERATIONS |
Whether batch operations are supported. | YES, NO |
SQL_CAP |
All supported SQL capabilities for this driver. | SELECT, INSERT, DELETE, UPDATE, TRANSACTIONS, ORDERBY, OAUTH, ASSIGNEDID, LIMIT, LIKE, BULKINSERT, COUNT, BULKDELETE, BULKUPDATE, GROUPBY, HAVING, AGGS, OFFSET, REPLICATE, COUNTDISTINCT, JOINS, DROP, CREATE, DISTINCT, INNERJOINS, SUBQUERIES, ALTER, MULTIPLESCHEMAS, GROUPBYNORELATION, OUTERJOINS, UNIONALL, UNION, UPSERT, GETDELETED, CROSSJOINS, GROUPBYCOLLATE, MULTIPLECATS, FULLOUTERJOIN, MERGE, JSONEXTRACT, BULKUPSERT, SUM, SUBQUERIESFULL, MIN, MAX, JOINSFULL, XMLEXTRACT, AVG, MULTISTATEMENTS, FOREIGNKEYS, CASE, LEFTJOINS, COMMAJOINS, WITH, LITERALS, RENAME, NESTEDTABLES, EXECUTE, BATCH, BASIC, INDEX |
PREFERRED_CACHE_OPTIONS |
A string value specifies the preferred cacheOptions. | |
ENABLE_EF_ADVANCED_QUERY |
Indicates if the driver directly supports advanced queries coming from Entity Framework. If not, queries will be handled client side. | YES, NO |
PSEUDO_COLUMNS |
A string array indicating the available pseudo columns. | |
MERGE_ALWAYS |
If the value is true, The Merge Mode is forcibly executed in Data Sync. | TRUE, FALSE |
REPLICATION_MIN_DATE_QUERY |
A select query to return the replicate start datetime. | |
REPLICATION_MIN_FUNCTION |
Allows a provider to specify the formula name to use for executing a server side min. | |
REPLICATION_START_DATE |
Allows a provider to specify a replicate startdate. | |
REPLICATION_MAX_DATE_QUERY |
A select query to return the replicate end datetime. | |
REPLICATION_MAX_FUNCTION |
Allows a provider to specify the formula name to use for executing a server side max. | |
IGNORE_INTERVALS_ON_INITIAL_REPLICATE |
A list of tables which will skip dividing the replicate into chunks on the initial replicate. | |
CHECKCACHE_USE_PARENTID |
Indicates whether the CheckCache statement should be done against the parent key column. | TRUE, FALSE |
CREATE_SCHEMA_PROCEDURES |
Indicates stored procedures that can be used for generating schema files. |
The following query retrieves the operators that can be used in the WHERE clause:
SELECT * FROM sys_sqlinfo WHERE Name = 'SUPPORTED_OPERATORS'
Note that individual tables may have different limitations or requirements on the WHERE clause; refer to the Data Model section for more information.
Columns
| Name | Type | Description |
|---|---|---|
NAME |
String |
A component of SQL syntax, or a capability that can be processed on the server. |
VALUE |
String |
Detail on the supported SQL or SQL syntax. |
sys_identity
Returns information about attempted modifications.
The following query retrieves the Ids of the modified rows in a batch operation:
SELECT * FROM sys_identity
Columns
| Name | Type | Description |
|---|---|---|
Id |
String |
The database-generated ID returned from a data modification operation. |
Batch |
String |
An identifier for the batch. 1 for a single operation. |
Operation |
String |
The result of the operation in the batch: INSERTED, UPDATED, or DELETED. |
Message |
String |
SUCCESS or an error message if the update in the batch failed. |
sys_information
Describes the available system information.
The following query retrieves all columns:
SELECT * FROM sys_information
Columns
| Name | Type | Description |
|---|---|---|
Product |
String |
The name of the product. |
Version |
String |
The version number of the product. |
Datasource |
String |
The name of the datasource the product connects to. |
NodeId |
String |
The unique identifier of the machine where the product is installed. |
HelpURL |
String |
The URL to the product's help documentation. |
License |
String |
The license information for the product. (If this information is not available, the field may be left blank or marked as 'N/A'.) |
Location |
String |
The file path location where the product's library is stored. |
Environment |
String |
The version of the environment or rumtine the product is currently running under. |
DataSyncVersion |
String |
The tier of Sync required to use this connector. |
DataSyncCategory |
String |
The category of Sync functionality (e.g., Source, Destination). |
Advanced Configurations Properties
The advanced configurations properties are the various options that can be used to establish a connection. This section provides a complete list of the options you can configure. Click the links for further details.
| Property | Description |
|---|---|
AuthScheme |
Specifies the authentication scheme the provider uses when connecting to Azure Data Lake Storage. |
Account |
This property specifies the name of the Azure Data Lake storage account. |
AccessKey |
Your Azure DataLakeStorage Gen 2 storage account access key. |
FileSystem |
Specifies the name of the file system in your Azure Data Lake Storage Gen2 account. |
SharedAccessSignature |
A shared access key signature that may be used for authentication. |
| Property | Description |
|---|---|
Directory |
Specifies the root directory path in Azure Data Lake Storage where the provider begins listing files and folders. |
IncludeSubDirectories |
Specifies whether the provider includes subdirectory paths in the Resources view. |
| Property | Description |
|---|---|
AzureTenant |
Identifies the Azure Data Lake Storage tenant being used to access data. Accepts either the tenant's domain name (for example, contoso.onmicrosoft.com) or its directory (tenant) ID. |
AzureEnvironment |
Specifies the Azure network environment to which you will connect. Must be the same network to which your Azure account was added. |
| Property | Description |
|---|---|
InitiateOAuth |
Specifies the process for obtaining or refreshing the OAuth access token, which maintains user access while an authenticated, authorized user is working. |
OAuthClientId |
Specifies the client ID (also known as the consumer key) assigned to your custom OAuth application. This ID is required to identify the application to the OAuth authorization server during authentication. |
OAuthClientSecret |
Specifies the client secret assigned to your custom OAuth application. This confidential value is used to authenticate the application to the OAuth authorization server. |
OAuthAccessToken |
Specifies the OAuth access token used to authenticate requests to the data source. This token is issued by the authorization server after a successful OAuth exchange. |
OAuthSettingsLocation |
Specifies the location of the settings file where OAuth values are saved. Storing OAuth settings in a central location avoids the need for users to enter OAuth connection properties manually each time they log in. It also enables credentials to be shared across connections or processes. |
CallbackURL |
Identifies the URL users return to after authenticating to Azure Data Lake Storage via OAuth. (Custom OAuth applications only.). |
OAuthVerifier |
Specifies a verifier code returned from the OAuthAuthorizationURL . Used when authenticating to OAuth on a headless server, where a browser can't be launched. Requires both OAuthSettingsLocation and OAuthVerifier to be set. |
OAuthRefreshToken |
Specifies the OAuth refresh token used to request a new access token after the original has expired. |
OAuthExpiresIn |
Specifies the duration in seconds, of an OAuth Access Token's lifetime. The token can be reissued to keep access alive as long as the user keeps working. |
OAuthTokenTimestamp |
Displays a Unix epoch timestamp in milliseconds that shows how long ago the current Access Token was created. |
| Property | Description |
|---|---|
OAuthJWTCert |
Supplies the name of the client certificate's JWT Certificate store. |
OAuthJWTCertType |
Identifies the type of key store containing the JWT Certificate. |
OAuthJWTCertPassword |
Provides the password for the OAuth JWT certificate used to access a password-protected certificate store. If the certificate store does not require a password, leave this property blank. |
OAuthJWTCertSubject |
Identifies the subject of the OAuth JWT certificate used to locate a matching certificate in the store. Supports partial matches and the wildcard '*' to select the first certificate. |
| Property | Description |
|---|---|
SSLServerCert |
Specifies the certificate to be accepted from the server when connecting using TLS/SSL. |
| Property | Description |
|---|---|
Location |
Specifies the location of a directory containing schema files that define tables, views, and stored procedures. Depending on your service's requirements, this may be expressed as either an absolute path or a relative path. |
BrowsableSchemas |
Optional setting that restricts the schemas reported to a subset of all available schemas. For example, BrowsableSchemas=SchemaA, SchemaB, SchemaC . |
Tables |
Optional setting that restricts the tables reported to a subset of all available tables. For example, Tables=TableA, TableB, TableC . |
Views |
Optional setting that restricts the views reported to a subset of the available tables. For example, Views=ViewA, ViewB, ViewC . |
| Property | Description |
|---|---|
ChunkSize |
The size of chunks (in MB) to use when uploading large files. |
MaxRows |
Specifies the maximum rows returned for queries without aggregation or GROUP BY. |
MaxThreads |
Specifies the number of concurrent requests. |
Other |
Specifies additional hidden properties for specific use cases. These are not required for typical provider functionality. Use a semicolon-separated list to define multiple properties. |
PseudoColumns |
Specifies the pseudocolumns to expose as table columns. Use the format 'TableName=ColumnName;TableName=ColumnName'. The default is an empty string, which disables this property. |
Timeout |
Specifies the maximum time, in seconds, that the provider waits for a server response before throwing a timeout error. The default is 60 seconds. Set to 0 to disable the timeout. |
UserDefinedViews |
Specifies a filepath to a JSON configuration file defining custom views. The provider automatically detects and uses the views specified in this file. |
Authentication
This section provides a complete list of authentication properties you can configure.
| Property | Description |
|---|---|
AuthScheme |
Specifies the authentication scheme the provider uses when connecting to Azure Data Lake Storage. |
Account |
This property specifies the name of the Azure Data Lake storage account. |
AccessKey |
Your Azure DataLakeStorage Gen 2 storage account access key. |
FileSystem |
Specifies the name of the file system in your Azure Data Lake Storage Gen2 account. |
SharedAccessSignature |
A shared access key signature that may be used for authentication. |
AuthScheme
Specifies the authentication scheme the provider uses when connecting to Azure Data Lake Storage.
Possible Values
Microsoft Entra ID, AzureServicePrincipal, AzureServicePrincipalCert, AzureMSI, AccessKey, SAS
Data Type
string
Default Value
Microsoft Entra ID
Remarks
If this property is left unset, the connector uses the default value: Microsoft Entra ID.
Microsoft Entra ID: Performs Microsoft Entra ID OAuth authentication.AzureServicePrincipal: Authenticates as a Microsoft Entra service principal (role-based or application-based) using a Client Secret.AzureServicePrincipalCert: Authenticates as a Microsoft Entra service principal (role-based or application-based) using a Certificate.AzureMSI: Automatically obtains Microsoft Entra ID Managed Service Identity credentials when running on an Azure VM.AccessKey: Authenticates the calls to the API by setting the AccessKey to your Access Key.SAS: Authenticates the calls to the API by setting the SharedAccessSignature to your Shared Access Signature (SAS).
Account
This property specifies the name of the Azure Data Lake storage account.
Data Type
string
Default Value
""
Remarks
The account name is the unique name assigned to your Azure Data Lake Storage Gen2 account. It is required for authentication and for constructing the service endpoint URL.
You can find the account name in the Azure portal:
- Sign in to
https://portal.azure.comwith your account credentials. - Navigate to
Storage accountsand select the account you want to use. - The
Namefield on theOverviewpage is your account name.
AccessKey
Your Azure DataLakeStorage Gen 2 storage account access key.
Data Type
string
Default Value
""
Remarks
The access key is one of the credentials used to authenticate requests to Azure Data Lake Storage Gen2 when AuthScheme is set to AccessKey.
You can retrieve the access key from the Azure portal:
- Sign in to
https://portal.azure.comwith your account credentials. - Navigate to
Storage accountsand select the account you want to use. - Under
Settings, selectAccess keys.
Keep your access key secure and avoid sharing it in logs or unencrypted configurations.
FileSystem
Specifies the name of the file system in your Azure Data Lake Storage Gen2 account.
Data Type
string
Default Value
""
Remarks
The file system is the top-level container for your data in Azure Data Lake Storage Gen2. It is equivalent to a container in Azure Blob Storage.
You can find the file system name in the Azure portal:
- Sign in to
https://portal.azure.comwith your account credentials. - Navigate to
Storage accountsand select your account. - Under
Data storage, selectContainers(file systems). - The
Namecolumn lists your available file systems.
SharedAccessSignature
A shared access key signature that may be used for authentication.
Data Type
string
Default Value
""
Remarks
A Shared Access Signature (SAS) token provides delegated access to resources in your Azure Data Lake Storage Gen2 account. You can use a SAS token instead of the AccessKey property to authenticate.
SAS tokens define the permissions granted (such as read, write, and list) and the time period for which they are valid.
You can generate a SAS token in the Azure portal:
- Sign in to
https://portal.azure.comwith your account credentials. - Navigate to
Storage accountsand select your account. - Under
Security + networking, selectShared access signature. - Choose the desired permissions and expiration date, then select
Generate SAS. - Copy the generated SAS token value into this property.
This property is only applicable when AuthScheme is set to SAS.
Connection
This section provides a complete list of connection properties you can configure.
| Property | Description |
|---|---|
Directory |
Specifies the root directory path in Azure Data Lake Storage where the provider begins listing files and folders. |
IncludeSubDirectories |
Specifies whether the provider includes subdirectory paths in the Resources view. |
Directory
Specifies the root directory path in Azure Data Lake Storage where the provider begins listing files and folders.
Data Type
string
Default Value
""
Remarks
The root directory path determines where in your Azure Data Lake Storage Gen2 file system the connector begins listing files and folders.
If this property is left unset, the connector lists files and folders from the root of the specified file system.
You can set this to a specific folder path to limit queries to a subset of the file system.
Directory paths are case-sensitive and must use forward slashes "/" as separators.
IncludeSubDirectories
Specifies whether the provider includes subdirectory paths in the Resources view.
Data Type
bool
Default Value
false
Remarks
This property determines whether the connector includes subdirectory paths when listing items in the Resources view.
If this property is set to true, the Resources view includes both the root-level items and any items in subdirectories.
If this property is set to false, only root-level items are listed, and subdirectories are omitted.
Use this setting to control whether listings should reflect the full hierarchy of files and folders or only top-level contents.
Azure Authentication
This section provides a complete list of Azure authentication properties you can configure.
| Property | Description |
|---|---|
AzureTenant |
Identifies the Azure Data Lake Storage tenant being used to access data. Accepts either the tenant's domain name (for example, contoso.onmicrosoft.com) or its directory (tenant) ID. |
AzureEnvironment |
Specifies the Azure network environment to which you will connect. Must be the same network to which your Azure account was added. |
AzureTenant
Identifies the Azure Data Lake Storage tenant being used to access data. Accepts either the tenant's domain name (for example, contoso.onmicrosoft.com) or its directory (tenant) ID.
Data Type
string
Default Value
""
Remarks
A tenant is a digital container for your organization's users and resources, managed through Microsoft Entra ID (formerly Microsoft Entra ID). Each tenant is associated with a unique directory ID, and often with a custom domain (for example, microsoft.com or contoso.onmicrosoft.com).
You can locate the directory (tenant) ID in the Microsoft Entra admin center by navigating to Microsoft Entra ID > Properties and copying the value labeled "Directory (tenant) ID".
This property is required in the following cases:
- When AuthScheme is set to
AzureServicePrincipalorAzureServicePrincipalCert - When AuthScheme is
Microsoft Entra IDand the user account belongs to multiple tenants
You can provide the tenant value in one of two formats:
- A domain name (for example, contoso.onmicrosoft.com)
- A directory (tenant) ID in GUID format (for example, c9d7b8e4-1234-4f90-bc1a-2a28e0f9e9e0)
Specifying the tenant explicitly ensures that the authentication request is routed to the correct directory, which is especially important when a user belongs to multiple tenants or when using service principal–based authentication.
If this value is omitted when required, authentication may fail or connect to the wrong tenant. This can result in errors such as unauthorized or resource not found.
AzureEnvironment
Specifies the Azure network environment to which you will connect. Must be the same network to which your Azure account was added.
Possible Values
GLOBAL, CHINA, USGOVT, USGOVTDOD
Data Type
string
Default Value
GLOBAL
Remarks
Required if your Azure account is part of a different network than the Global network, such as China, USGOVT, or USGOVTDOD.
OAuth
This section provides a complete list of OAuth properties you can configure.
| Property | Description |
|---|---|
InitiateOAuth |
Specifies the process for obtaining or refreshing the OAuth access token, which maintains user access while an authenticated, authorized user is working. |
OAuthClientId |
Specifies the client ID (also known as the consumer key) assigned to your custom OAuth application. This ID is required to identify the application to the OAuth authorization server during authentication. |
OAuthClientSecret |
Specifies the client secret assigned to your custom OAuth application. This confidential value is used to authenticate the application to the OAuth authorization server. |
OAuthAccessToken |
Specifies the OAuth access token used to authenticate requests to the data source. This token is issued by the authorization server after a successful OAuth exchange. |
OAuthSettingsLocation |
Specifies the location of the settings file where OAuth values are saved. Storing OAuth settings in a central location avoids the need for users to enter OAuth connection properties manually each time they log in. It also enables credentials to be shared across connections or processes. |
CallbackURL |
Identifies the URL users return to after authenticating to Azure Data Lake Storage via OAuth. (Custom OAuth applications only.). |
OAuthVerifier |
Specifies a verifier code returned from the OAuthAuthorizationURL . Used when authenticating to OAuth on a headless server, where a browser can't be launched. Requires both OAuthSettingsLocation and OAuthVerifier to be set. |
OAuthRefreshToken |
Specifies the OAuth refresh token used to request a new access token after the original has expired. |
OAuthExpiresIn |
Specifies the duration in seconds, of an OAuth Access Token's lifetime. The token can be reissued to keep access alive as long as the user keeps working. |
OAuthTokenTimestamp |
Displays a Unix epoch timestamp in milliseconds that shows how long ago the current Access Token was created. |
InitiateOAuth
Specifies the process for obtaining or refreshing the OAuth access token, which maintains user access while an authenticated, authorized user is working.
Possible Values
OFF, REFRESH, GETANDREFRESH
Data Type
string
Default Value
OFF
Remarks
OAuth is an authorization framework that enables applications to obtain limited access to user accounts on an HTTP service. The OAuth flow defines the method to be used for logging in users, exchanging their credentials for an OAuth access token to be used for authentication, and providing limited access to applications.
Azure Data Lake Storage supports the following options for initiating OAuth access:
OFF: No automatic OAuth flow initiation. The OAuth flow is handled entirely by the user, who will take action to obtain their OAuthAccessToken. Note that with this setting the user must refresh the token manually and reconnect with an updated OAuthAccessToken property when the current token expires.GETANDREFRESH: The OAuth flow is handled entirely by the connector. If a token already exists, it is refreshed when necessary. If no token currently exists, it will be obtained by prompting the user to login.REFRESH: The user handles obtaining the OAuth Access Token and sets up the sequence for refreshing the OAuth Access Token. (The user is never prompted to log in to authenticate. After the user logs in, the connector handles the refresh of the OAuth Access Token.
OAuthClientId
Specifies the client ID (also known as the consumer key) assigned to your custom OAuth application. This ID is required to identify the application to the OAuth authorization server during authentication.
Data Type
string
Default Value
""
Remarks
This property is required when using a custom OAuth application, such as in web-based authentication flows, service-based authentication, or certificate-based flows that require application registration. It is also required if an embedded OAuth application is not available for the driver. When an embedded OAuth application is available, this value may already be provided by the connector and not require manual entry.
This value is generally used alongside other OAuth-related properties such as OAuthClientSecret and OAuthSettingsLocation when configuring an authenticated connection.
OAuthClientId is one of the key connection parameters that need to be set before users can authenticate via OAuth. You can typically find this value in your identity provider’s application registration settings. Look for a field labeled Client ID, Application ID, or Consumer Key.
While the client ID is not considered a confidential value like a client secret, it is still part of your application's identity and should be handled carefully. Avoid exposing it in public repositories or shared configuration files.
OAuthClientSecret
Specifies the client secret assigned to your custom OAuth application. This confidential value is used to authenticate the application to the OAuth authorization server.
Data Type
string
Default Value
""
Remarks
This property is required when using a custom OAuth application in any flow that requires secure client authentication, such as web-based OAuth, service-based connections, or certificate-based authorization flows. It is not required when using an embedded OAuth application.
The client secret is used during the token exchange step of the OAuth flow, when the driver requests an access token from the authorization server. If this value is missing or incorrect, authentication will fail, and the server may return an invalid_client or unauthorized_client error.
OAuthClientSecret is one of the key connection parameters that need to be set before users can authenticate via OAuth. You can obtain this value from your identity provider when registering the OAuth application. It may be referred to as the client secret, application secret, or consumer secret.
This value should be stored securely and never exposed in public repositories, scripts, or unsecured environments. Client secrets may also expire after a set period. Be sure to monitor expiration dates and rotate secrets as needed to maintain uninterrupted access.
OAuthAccessToken
Specifies the OAuth access token used to authenticate requests to the data source. This token is issued by the authorization server after a successful OAuth exchange.
Data Type
string
Default Value
""
Remarks
The OAuthAccessToken is a temporary credential that authorizes access to protected resources. It is typically returned by the identity provider after the user or client application completes an OAuth authentication flow. This property is most commonly used in automated workflows or custom OAuth implementations where you want to manage token handling outside of the driver.
The OAuth access token has a server-dependent timeout, limiting user access. This is set using the OAuthExpiresIn property. However, it can be reissued between requests to keep access alive as long as the user keeps working.
If InitiateOAuth is set to REFRESH, we recommend that you also set both OAuthExpiresIn and OAuthTokenTimestamp. The connector uses these properties to determine when the token expires so it can refresh most efficiently. If OAuthExpiresIn and OAuthTokenTimestamp are not specified, the connector refreshes the token immediately.
Access tokens should be treated as sensitive credentials and stored securely. Avoid exposing them in logs, scripts, or configuration files that are not access-controlled.
OAuthSettingsLocation
The location of the settings file where OAuth values are saved when InitiateOAuth is set to GETANDREFRESH or REFRESH. Alternatively, you can hold this location in memory by specifying a value starting with 'memory://'.
Data Type
string
Default Value
%APPDATA%\\CData\\Acumatica Data Provider\\OAuthSettings.txt
Remarks
When InitiateOAuth is set to GETANDREFRESH or REFRESH, the driver saves OAuth values to avoid requiring the user to manually enter OAuth connection properties and to allow the credentials to be shared across connections or processes.
Instead of specifying a file path, you can use memory storage. Memory locations are specified by using a value starting with 'memory://' followed by a unique identifier for that set of credentials (for example, memory://user1). The identifier can be anything you choose but should be unique to the user. Unlike file-based storage, where credentials persist across connections, memory storage loads the credentials into static memory, and the credentials are shared between connections using the same identifier for the life of the process. To persist credentials outside the current process, you must manually store the credentials prior to closing the connection. This enables you to set them in the connection when the process is started again. You can retrieve OAuth property values with a query to the sys_connection_props system table. If there are multiple connections using the same credentials, the properties are read from the previously closed connection.
The default location is "%APPDATA%\\CData\\Acumatica Data Provider\\OAuthSettings.txt" with %APPDATA% set to the user's configuration directory. The default values are
- Windows: "
register://%DSN" - Unix: "%AppData%..."
where DSN is the name of the current DSN used in the open connection.
The following table lists the value of %APPDATA% by OS:
| Platform | %APPDATA% |
|---|---|
Windows |
The value of the APPDATA environment variable |
Linux |
~/.config |
CallbackURL
Identifies the URL users return to after authenticating to Azure Data Lake Storage via OAuth. (Custom OAuth applications only.).
Data Type
string
Default Value
""
Remarks
If you created a custom OAuth application, the OAuth authorization server redirects the user to this URL during the authentication process. This value must match the callback URL you specified when you Configured the custom OAuth application.
OAuthVerifier
Specifies a verifier code returned from the OAuthAuthorizationURL . Used when authenticating to OAuth on a headless server, where a browser can't be launched. Requires both OAuthSettingsLocation and OAuthVerifier to be set.
Data Type
string
Default Value
""
OAuthRefreshToken
Specifies the OAuth refresh token used to request a new access token after the original has expired.
Data Type
string
Default Value
""
Remarks
The refresh token is used to obtain a new access token when the current one expires. It enables seamless authentication for long-running or automated workflows without requiring the user to log in again. This property is especially important in headless, CI/CD, or server-based environments where interactive authentication is not possible.
The refresh token is typically obtained during the initial OAuth exchange by calling the GetOAuthAccessToken stored procedure. After that, it can be set using this property to enable automatic token refresh, or passed to the RefreshOAuthAccessToken stored procedure if you prefer to manage the refresh manually.
When InitiateOAuth is set to REFRESH, the driver uses this token to retrieve a new access token automatically. After the first refresh, the driver saves updated tokens in the location defined by OAuthSettingsLocation, and uses those values for subsequent connections.
The OAuthRefreshToken should be handled securely and stored in a trusted location. Like access tokens, refresh tokens can expire or be revoked depending on the identity provider’s policies.
OAuthExpiresIn
Specifies the duration in seconds, of an OAuth Access Token's lifetime. The token can be reissued to keep access alive as long as the user keeps working.
Data Type
string
Default Value
""
Remarks
The OAuth Access Token is assigned to an authenticated user, granting that user access to the network for a specified period of time. The access token is used in place of the user's login ID and password, which stay on the server.
An access token created by the server is only valid for a limited time. OAuthExpiresIn is the number of seconds the token is valid from when it was created. For example, a token generated at 2024-01-29 20:00:00 UTC that expires at 2024-01-29 21:00:00 UTC (an hour later) would have an OAuthExpiresIn value of 3600, no matter what the current time is.
To determine how long the user has before the Access Token will expire, use OAuthTokenTimestamp.
OAuthTokenTimestamp
Displays a Unix epoch timestamp in milliseconds that shows how long ago the current Access Token was created.
Data Type
string
Default Value
""
Remarks
The OAuth Access Token is assigned to an authenticated user, granting that user access to the network for a specified period of time. The access token is used in place of the user's login ID and password, which stay on the server.
An access token created by the server is only valid for a limited time. OAuthTokenTimestamp is the Unix timestamp when the server created the token. For example, OAuthTokenTimestamp=1706558400 indicates the OAuthAccessToken was generated by the server at 2024-01-29 20:00:00 UTC.
JWT OAuth
This section provides a complete list of JWT OAuth properties you can configure.
| Property | Description |
|---|---|
OAuthJWTCert |
Supplies the name of the client certificate's JWT Certificate store. |
OAuthJWTCertType |
Identifies the type of key store containing the JWT Certificate. |
OAuthJWTCertPassword |
Provides the password for the OAuth JWT certificate used to access a password-protected certificate store. If the certificate store does not require a password, leave this property blank. |
OAuthJWTCertSubject |
Identifies the subject of the OAuth JWT certificate used to locate a matching certificate in the store. Supports partial matches and the wildcard '*' to select the first certificate. |
OAuthJWTCert
Supplies the name of the client certificate's JWT Certificate store.
Data Type
string
Default Value
""
Remarks
The OAuthJWTCertType field specifies the type of the certificate store specified in OAuthJWTCert. If the store is password-protected, use OAuthJWTCertPassword to supply the password..
OAuthJWTCert is used in conjunction with the OAuthJWTCertSubject field in order to specify client certificates.
If OAuthJWTCert has a value, and OAuthJWTCertSubject is set, the Azure Data Lake Storage connector initiates a search for a certificate. For further information, see OAuthJWTCertSubject.
Designations of certificate stores are platform-dependent.
Notes
-
The most common User and Machine certificate stores in Windows include:
-
MY: A certificate store holding personal certificates with their
associated private keys.
-
CA: Certifying authority certificates.
-
ROOT: Root certificates.
-
SPC: Software publisher certificates.
- In Java, the certificate store normally is a file containing certificates and optional private keys.
- When the certificate store type is PFXFile, this property must be set to the name of the file.
- When the type is PFXBlob, the property must be set to the binary contents of a PFX file (i.e. PKCS12 certificate store).
-
OAuthJWTCertType
Identifies the type of key store containing the JWT Certificate.
Possible Values
USER, MACHINE, PFXFILE, PFXBLOB, JKSFILE, JKSBLOB, PEMKEY_FILE, PEMKEY_BLOB, PUBLIC_KEY_FILE, PUBLIC_KEY_BLOB, SSHPUBLIC_KEY_FILE, SSHPUBLIC_KEY_BLOB, P7BFILE, PPKFILE, XMLFILE, XMLBLOB, BCFKSFILE, BCFKSBLOB
Data Type
string
Default Value
USER
Remarks
| Value | Description | Notes |
|---|---|---|
| USER | A certificate store owned by the current user. | Only available in Windows. |
| MACHINE | A machine store. | Not available in Java or other non-Windows environments. |
| PFXFILE | A PFX (PKCS12) file containing certificates. | |
| PFXBLOB | A string (base-64-encoded) representing a certificate store in PFX (PKCS12) format. | |
| JKSFILE | A Java key store (JKS) file containing certificates. | Only available in Java. |
| JKSBLOB | A string (base-64-encoded) representing a certificate store in Java key store (JKS) format. | Only available in Java. |
| PEMKEY_FILE | A PEM-encoded file that contains a private key and an optional certificate. | |
| PEMKEY_BLOB | A string (base64-encoded) that contains a private key and an optional certificate. | |
| PUBLIC_KEY_FILE | A file that contains a PEM- or DER-encoded public key certificate. | |
| PUBLIC_KEY_BLOB | A string (base-64-encoded) that contains a PEM- or DER-encoded public key certificate. | |
| SSHPUBLIC_KEY_FILE | A file that contains an SSH-style public key. | |
| SSHPUBLIC_KEY_BLOB | A string (base-64-encoded) that contains an SSH-style public key. | |
| P7BFILE | A PKCS7 file containing certificates. | |
| PPKFILE | A file that contains a PPK (PuTTY Private Key). | |
| XMLFILE | A file that contains a certificate in XML format. | |
| XMLBLOB | Astring that contains a certificate in XML format. | |
| BCFKSFILE | A file that contains an Bouncy Castle keystore. | |
| BCFKSBLOB | A string (base-64-encoded) that contains a Bouncy Castle keystore. |
OAuthJWTCertPassword
Provides the password for the OAuth JWT certificate used to access a password-protected certificate store. If the certificate store does not require a password, leave this property blank.
Data Type
string
Default Value
""
Remarks
This property specifies the password needed to open a password-protected certificate store. To determine if a password is necessary, refer to the documentation or configuration for your specific certificate store.
OAuthJWTCertSubject
Identifies the subject of the OAuth JWT certificate used to locate a matching certificate in the store. Supports partial matches and the wildcard '*' to select the first certificate.
Data Type
string
Default Value
*
Remarks
The value of this property is used to locate a matching certificate in the store. The search process works as follows:
- If an exact match for the subject is found, the corresponding certificate is selected.
- If no exact match is found, the store is searched for certificates whose subjects contain the property value.
- If no match is found, no certificate is selected.
You can set the value to '*' to automatically select the first certificate in the store. The certificate subject is a comma-separated list of distinguished name fields and values. For example: CN=www.server.com, OU=test, C=US, E=example@jbexample.com. Common fields include:
| Field | Meaning |
|---|---|
CN |
Common Name. This is commonly a host name like www.server.com. |
O |
Organization |
OU |
Organizational Unit |
L |
Locality |
S |
State |
C |
Country |
E |
Email Address |
If a field value contains a comma, enclose it in quotes. For example: "O=ACME, Inc.".
SSL
This section provides a complete list of SSL properties you can configure.
| Property | Description |
|---|---|
SSLServerCert |
Specifies the certificate to be accepted from the server when connecting using TLS/SSL. |
SSLServerCert
Specifies the certificate to be accepted from the server when connecting using TLS/SSL.
Data Type
string
Default Value
""
Remarks
If using a TLS/SSL connection, this property can be used to specify the TLS/SSL certificate to be accepted from the server. Any other certificate that is not trusted by the machine is rejected.
This property can take the following forms:
| Description | Example |
|---|---|
| A full PEM Certificate (example shortened for brevity) | -----BEGIN CERTIFICATE----- MIIChTCCAe4CAQAwDQYJKoZIhv......Qw== -----END CERTIFICATE----- |
| A path to a local file containing the certificate | C:\\cert.cer |
| The public key (example shortened for brevity) | -----BEGIN RSA PUBLIC KEY----- MIGfMA0GCSq......AQAB -----END RSA PUBLIC KEY----- |
| The MD5 Thumbprint (hex values can also be either space or colon separated) | ecadbdda5a1529c58a1e9e09828d70e4 |
| The SHA1 Thumbprint (hex values can also be either space or colon separated) | 34a929226ae0819f2ec14b4a3d904f801cbb150d |
If not specified, any certificate trusted by the machine is accepted.
Certificates are validated as trusted by the machine based on the System's trust store. The trust store used is the 'javax.net.ssl.trustStore' value specified for the system. If no value is specified for this property, Java's default trust store is used (for example, JAVA_HOME\lib\security\cacerts).
Use '*' to signify to accept all certificates. Note that this is not recommended due to security concerns.
Schema
This section provides a complete list of schema properties you can configure.
| Property | Description |
|---|---|
Location |
Specifies the location of a directory containing schema files that define tables, views, and stored procedures. Depending on your service's requirements, this may be expressed as either an absolute path or a relative path. |
BrowsableSchemas |
Optional setting that restricts the schemas reported to a subset of all available schemas. For example, BrowsableSchemas=SchemaA, SchemaB, SchemaC . |
Tables |
Optional setting that restricts the tables reported to a subset of all available tables. For example, Tables=TableA, TableB, TableC . |
Views |
Optional setting that restricts the views reported to a subset of the available tables. For example, Views=ViewA, ViewB, ViewC . |
Location
Specifies the location of a directory containing schema files that define tables, views, and stored procedures. Depending on your service's requirements, this may be expressed as either an absolute path or a relative path.
Data Type
string
Default Value
%APPDATA%\ADLS Data Provider\Schema
Remarks
The Location property is only needed if you want to either customize definitions (for example, change a column name, ignore a column, etc.) or extend the data model with new tables, views, or stored procedures.
Note
Since this connector supports multiple schemas, custom schema files for Azure Data Lake Storage should be structured such that:
- Each schema should have its own folder, named for that schema.
- All schema folders should be contained in a parent folder.
Location should always be set to the parent folder, and not to an individual schema's folder.
If left unspecified, the default location is %APPDATA%\ADLS Data Provider\Schema, where %APPDATA% is set to the user's configuration directory:
| Platform | %APPDATA% |
|---|---|
Windows |
The value of the APPDATA environment variable |
Mac |
~/Library/Application Support |
Linux |
~/.config |
BrowsableSchemas
Optional setting that restricts the schemas reported to a subset of all available schemas. For example, BrowsableSchemas=SchemaA,SchemaB,SchemaC.
Data Type
string
Default Value
""
Remarks
Listing all available database schemas can take extra time, thus degrading performance. Providing a list of schemas in the connection string saves time and improves performance.
Tables
Optional setting that restricts the tables reported to a subset of all available tables. For example, Tables=TableA,TableB,TableC.
Data Type
string
Default Value
""
Remarks
Listing all available tables from some databases can take extra time, thus degrading performance. Providing a list of tables in the connection string saves time and improves performance.
If there are lots of tables available and you already know which ones you want to work with, you can use this property to restrict your viewing to only those tables. To do this, specify the tables you want in a comma-separated list. Each table should be a valid SQL identifier with any special characters escaped using square brackets, double-quotes or backticks. For example, Tables=TableA,[TableB/WithSlash],WithCatalog.WithSchema.`TableC With Space`.
Note
If you are connecting to a data source with multiple schemas or catalogs, you must specify each table you want to view by its fully qualified name. This avoids ambiguity between tables that may exist in multiple catalogs or schemas.
Views
Optional setting that restricts the views reported to a subset of the available tables. For example, Views=ViewA,ViewB,ViewC.
Data Type
string
Default Value
""
Remarks
Listing all available views from some databases can take extra time, thus degrading performance. Providing a list of views in the connection string saves time and improves performance.
If there are lots of views available and you already know which ones you want to work with, you can use this property to restrict your viewing to only those views. To do this, specify the views you want in a comma-separated list. Each view should be a valid SQL identifier with any special characters escaped using square brackets, double-quotes or backticks. For example, Views=ViewA,[ViewB/WithSlash],WithCatalog.WithSchema.`ViewC With Space`.
Note
If you are connecting to a data source with multiple schemas or catalogs, you must specify each view you want to examine by its fully qualified name. This avoids ambiguity between views that may exist in multiple catalogs or schemas.
Miscellaneous
This section provides a complete list of miscellaneous properties you can configure.
| Property | Description |
|---|---|
ChunkSize |
The size of chunks (in MB) to use when uploading large files. |
MaxRows |
Specifies the maximum rows returned for queries without aggregation or GROUP BY. |
MaxThreads |
Specifies the number of concurrent requests. |
Other |
Specifies additional hidden properties for specific use cases. These are not required for typical provider functionality. Use a semicolon-separated list to define multiple properties. |
PseudoColumns |
Specifies the pseudocolumns to expose as table columns. Use the format 'TableName=ColumnName;TableName=ColumnName'. The default is an empty string, which disables this property. |
Timeout |
Specifies the maximum time, in seconds, that the provider waits for a server response before throwing a timeout error. The default is 60 seconds. Set to 0 to disable the timeout. |
UserDefinedViews |
Specifies a filepath to a JSON configuration file defining custom views. The provider automatically detects and uses the views specified in this file. |
ChunkSize
The size of chunks (in MB) to use when uploading large files.
Data Type
int
Default Value
16
Remarks
This property determines the size, in megabytes, of each data block the connector uploads when transferring large files to Azure Data Lake Storage Gen2.
If this property is left unset, the connector uses the default value: 16 MB.
If a file exceeds the chunk size, it is divided into multiple blocks of the specified size and uploaded sequentially.
Increasing the chunk size can improve upload performance for large files, but may increase memory usage. Decreasing it can reduce memory use, but may result in slower uploads.
This setting is only applied to upload operations and does not affect file downloads.
MaxRows
Specifies the maximum rows returned for queries without aggregation or GROUP BY.
Data Type
int
Default Value
-1
Remarks
This property sets an upper limit on the number of rows the connector returns for queries that do not include aggregation or GROUP BY clauses. This limit ensures that queries do not return excessively large result sets by default.
When a query includes a LIMIT clause, the value specified in the query takes precedence over the MaxRows setting. If MaxRows is set to "-1", no row limit is enforced unless a LIMIT clause is explicitly included in the query.
This property is useful for optimizing performance and preventing excessive resource consumption when executing queries that could otherwise return very large datasets.
MaxThreads
Specifies the number of concurrent requests.
Data Type
string
Default Value
5
Remarks
This property allows you to issue multiple requests simultaneously, thereby improving performance.
Other
Specifies additional hidden properties for specific use cases. These are not required for typical provider functionality. Use a semicolon-separated list to define multiple properties.
Data Type
string
Default Value
""
Remarks
This property allows advanced users to configure hidden properties for specialized scenarios. These settings are not required for normal use cases but can address unique requirements or provide additional functionality. Multiple properties can be defined in a semicolon-separated list.
Note
It is strongly recommended to set these properties only when advised by the support team to address specific scenarios or issues.
Specify multiple properties in a semicolon-separated list.
Integration and Formatting
| Property | Description |
|---|---|
DefaultColumnSize |
Sets the default length of string fields when the data source does not provide column length in the metadata. The default value is 2000. |
ConvertDateTimeToGMT=True |
Converts date-time values to GMT, instead of the local time of the machine. The default value is False (use local time). |
RecordToFile=filename |
Records the underlying socket data transfer to the specified file. |
PseudoColumns
Specifies the pseudocolumns to expose as table columns. Use the format 'TableName=ColumnName;TableName=ColumnName'. The default is an empty string, which disables this property.
Data Type
string
Default Value
""
Remarks
This property allows you to define which pseudocolumns the connector exposes as table columns.
To specify individual pseudocolumns, use the following format: "Table1=Column1;Table1=Column2;Table2=Column3"
To include all pseudocolumns for all tables use: "*=*"
Timeout
Specifies the maximum time, in seconds, that the provider waits for a server response before throwing a timeout error. The default is 60 seconds. Set to 0 to disable the timeout.
Data Type
int
Default Value
60
Remarks
This property controls the maximum time, in seconds, that the connector waits for an operation to complete before canceling it. If the timeout period expires before the operation finishes, the connector cancels the operation and throws an exception.
The timeout applies to each individual communication with the server rather than the entire query or operation. For example, a query could continue running beyond the timeout value if each paging call completes within the timeout limit.
Setting this property to 0 disables the timeout, allowing operations to run indefinitely until they succeed or fail due to other conditions such as server-side timeouts, network interruptions, or resource limits on the server. Use this property cautiously to avoid long-running operations that could degrade performance or result in unresponsive behavior.
UserDefinedViews
Specifies a filepath to a JSON configuration file defining custom views. The provider automatically detects and uses the views specified in this file.
Data Type
string
Default Value
""
Remarks
This property allows you to define and manage custom views through a JSON-formatted configuration file called UserDefinedViews.json. These views are automatically recognized by the connector and enable you to execute custom SQL queries as if they were standard database views. The JSON file defines each view as a root element with a child element called "query", which contains the SQL query for the view. For example:
{
"MyView": {
"query": "SELECT * FROM Resources WHERE MyColumn = 'value'"
},
"MyView2": {
"query": "SELECT * FROM MyTable WHERE Id IN (1,2,3)"
}
}
You can define multiple views in a single file and specify the filepath using this property. For example: UserDefinedViews=C:\Path\To\UserDefinedViews.json. When you use this property, only the specified views are seen by the connector.
Refer to User Defined Views for more information.