Explore data
Last update: February 14, 2025
- Topics:
- Queries
CREATED FOR:
- Beginner
- Developer
Learn how to validate ingested data, preview data, and explore statistical and analytical properties of data using SQL functions. For more information, please visit the Query Service documentation.

Transcript
Hi there. Adobe Experience Platform retrieves data from a wide variety of sources. Immediate challenge for marketers is making sense of this data to gain insights about their customers Adobe Experience Platform Query Service facilitates that by allowing you to use standard SQL to query data in platform using in user interface. In this video, let me show you how data engineers can quickly explore and validate data industry to datasets. From your Experience Platform homepage, navigate your datasets under data management. For this video, you’re using a fictional retail brand called Luma. Luma Loyalty dataset contains information about customers, loyalty details, geographic details, et cetera. Luma web data contains web traffic information about customers interaction on the Luma site, including products viewed by a customer, visited pages, products purchased, et cetera. Open the Luma Loyalty dataset and the last successful bachelor had ingested over 5,000 customer data. Now let’s verify the industry records by clicking on the preview dataset option. Please take a closer look at the records and their structure. As a data engineer, you can verify what’s stored in a dataset using the “Previous dataset” option. However, the preview option displays only a few records, or 200 records, but that’s not the entire dataset. Some records might have incorrect values are missing data. So to make sure that the data ingested into a dataset contains the expected number of records to explore the data type and format for each column, we can use Experience Platform query service UI, or any Postgres SQL compliant tool of your choice. From the left navigation bar, click on “credits” under “data management.” Let’s click on the “Create Query” , and you can see a query editor window open up. Let’s make sure that the number of records ingested into the dataset is the same as the source file. In the resell stamp, you can see that we have 5,000 records in the dataset, which matches the record entered into the dataset. So we made sure that no data is missing. Now let’s the dataset to make sure there are no duplicate customer loyalty records in the dataset. To do that, let’s write the query to output the unique records in the Luma Loyalty dataset. After 5,000 records, we can now confirm that there are a few duplicate records in the dataset. Let’s run a query to output duplicate records of loyalty numbers in the dataset. Using the result generated, you can then remove the duplicate records from the dataset, or write a query to output the dataset that doesn’t contain any duplicate records. To perform additional validation, Adobe Experience Platform Query Service provides several built-in Spark SQL functions to extend SQL functionality. Let’s see how we can use spark SQL functions to write a query, to check that there are no loyalty members who have a birthdate before 1900. To do this, I will be using the “Year” functionality under “date, time functions.” The date function returns the year when fed with a date. Let’s fish through the "Query Editor, and execute our query. After the query executes, you can notice that there are no records returned for customers who have a birth year less than 1900. This means we have a valid data ingested into the dataset and is ready to be used. Similarly, you can check out the platform query service document base to explore other Spark SQL functions that you could use to validate data. You can also connect to query service using different EXROP applications and tools that support cost rescuer to write and run your queries. Using 20 servers in Adobe Experience Platform that our engineers can quickly explore and validate data industry to datasets. -
Previous pageQuery Service API
Next pagePrepare Data
Experience Platform
- Platform Tutorials
- Introduction to Platform
- A customer experience powered by Experience Platform
- Behind the scenes: A customer experience powered by Experience Platform
- Experience Platform overview
- Key capabilities
- Platform-based applications
- Integrations with Experience Cloud applications
- Key use cases
- Basic architecture
- User interface
- Roles and project phases
- Introduction to Real-Time CDP
- Getting started: Data Architects and Data Engineers
- Authenticate to Experience Platform APIs
- Import sample data to Experience Platform
- Administration
- AI Assistant
- Audiences and Segmentation
- Introduction to Audience Portal and Composition
- Upload audiences
- Overview of Federated Audience Composition
- Connect and configure Federated Audience Composition
- Create a Federated Audience Composition
- Audience rule builder overview
- Create audiences
- Use time constraints
- Create content-based audiences
- Create conversion audiences
- Create audiences from existing audiences
- Create sequential audiences
- Create dynamic audiences
- Create multi-entity audiences
- Create and activate account audiences (B2B)
- Demo of streaming segmentation
- Evaluate batch audiences on demand
- Evaluate an audience rule
- Create a dataset to export data
- Segment Match connection setup
- Segment Match data governance
- Segment Match configuration flow
- Segment Match pre-share insights
- Segment Match receiving data
- Audit logs
- Data Collection
- Collaboration
- Dashboards
- Data Governance
- Data Hygiene
- Data Ingestion
- Overview
- Batch ingestion overview
- Create and populate a dataset
- Delete datasets and batches
- Map a CSV file to XDM
- Sources overview
- Ingest data from Adobe Analytics
- Ingest data from Audience Manager
- Ingest data from cloud storage
- Ingest data from CRM
- Ingest data from databases
- Streaming ingestion overview
- Stream data with HTTP API
- Stream data using Source Connectors
- Web SDK tutorials
- Mobile SDK tutorials
- Data Lifecycle
- Destinations
- Destinations overview
- Connect to destinations
- Create destinations and activate data
- Activate profiles and audiences to a destination
- Export datasets using a cloud storage destination
- Integrate with Google Customer Match
- Configure the Azure Blob destination
- Configure the Marketo destination
- Configure file-based cloud storage or email marketing destinations
- Configure a social destination
- Activate through LiveRamp destinations
- Adobe Target and Custom Personalization
- Activate data to non-Adobe applications webinar
- Identities
- Intelligent Services
- Monitoring
- Partner data support
- Profiles
- Understanding Real-Time Customer Profile
- Profile overview diagram
- Bring data into Profile
- Customize profile view details
- View account profiles
- Create merge policies
- Union schemas overview
- Create a computed attribute
- Pseudonymous profile expirations (TTL)
- Delete profiles
- Update a specific attribute using upsert
- Privacy and Security
- Introduction to Privacy Service
- Identity data in Privacy requests
- Privacy JavaScript library
- Privacy labels in Adobe Analytics
- Getting started with the Privacy Service API
- Privacy Service UI
- Privacy Service API
- Subscribe to Privacy Events
- Set up customer-managed keys
- 10 considerations for Responsible Customer Data Management
- Elevating the Marketer’s Role as a Data Steward
- Queries
- Overview
- Query Service UI
- Query Service API
- Explore Data
- Prepare Data
- Adobe Defined Functions
- Data usage patterns
- Run queries
- Generate datasets from query results
- Tableau
- Analyze and visualize data
- Build dashboards using BI tools
- Recharge your customer data
- Connect clients to Query Service
- Validate data in the datalake
- Schemas
- Overview
- Building blocks
- Plan your data model
- Convert your data model to XDM
- Create schemas
- Create schemas for B2B data
- Create classes
- Create field groups
- Create data types
- Configure relationships between schemas
- Use enumerated fields and suggested values
- Copy schemas between sandboxes
- Update schemas
- Create an ad hoc schema
- Sources
- Use Case Playbooks
- Experience Cloud Integrations
- Industry Trends