Read this page to learn to create a live outbound connection to your Azure Data Lake Storage Gen2 (ADLS Gen2) data lake to periodically export data files from Experience Platform.
This section describes which type of audiences you can export to this destination.
|Segmentation Service||✓||Audiences generated through the Experience Platform Segmentation Service.|
|Custom uploads||✓||Audiences imported into Experience Platform from CSV files.|
Refer to the table below for information about the destination export type and frequency.
|Export type||Profile-based||You are exporting all members of a segment, together with the applicable schema fields (for example your PPID), as chosen in the select profile attributes screen of the destination activation workflow.|
|Export frequency||Batch||Batch destinations export files to downstream platforms in increments of three, six, eight, twelve, or twenty-four hours. Read more about batch file-based destinations.|
To connect to this destination, follow the steps described in the destination configuration tutorial. In the destination configuration workflow, fill in the fields listed in the two sections below.
To authenticate to the destination, fill in the required fields and select Connect to destination.
URL: The endpoint for Azure Data Lake Storage Gen2. The endpoint pattern is:
Tenant: The tenant information that contains your application.
Service principal ID: The application’s client ID.
Service principal key: The application’s key.
Encryption key: Optionally, you can attach your RSA-formatted public key to add encryption to your exported files. View an example of a correctly formatted encryption key in the image below.
To configure details for the destination, fill in the required and optional fields below. An asterisk next to a field in the UI indicates that the field is required.
manifest-<<destinationId>>-<<dataflowRunId>>.json. View a sample manifest file. The manifest file includes the following fields:
flowRunId: The dataflow run which generated the exported file.
scheduledTime: The time in UTC when the file was exported.
exportResults.sinkPath: The path in your storage location where the exported file is deposited.
exportResults.name: The name of the exported file.
size: The size of the exported file, in bytes.
You can enable alerts to receive notifications on the status of the dataflow to your destination. Select an alert from the list to subscribe to receive notifications on the status of your dataflow. For more information on alerts, see the guide on subscribing to destinations alerts using the UI.
When you are finished providing details for your destination connection, select Next.
See Activate audience data to batch profile export destinations for instructions on activating audiences to this destination.
In the Mapping step, you can select which attribute and identity fields to export for your profiles. You can also select to change the headers in the exported file to any friendly name that you wish. For more information, view the mapping step in the activate batch destinations UI tutorial.
This destination supports dataset exports. For complete information on how to set up dataset exports, read the tutorials:
To verify if data has been exported successfully, check your Azure Data Lake Storage Gen2 storage and make sure that the exported files contain the expected profile populations.