Example use case for ÃÛ¶¹ÊÓƵ Experience Platform Query Service
This document and accompanying video presentation provide a high-level end-to-end workflow demonstrating how ÃÛ¶¹ÊÓƵ Experience Platform Query Service benefits your organization’s strategic business insights. Using a browse abandonment use case as an example, this guide illustrates the following key concepts:
- The key importance of data processing to maximize the potential of ÃÛ¶¹ÊÓƵ Experience Platform.
- Ways to build the query based on your existing data architecture.
- Ensure data quality that meets your needs, and methods to mitigate any shortfalls.
- The process to schedule a query to run at a set frequency for use downstream in segmentation and destinations for personalization.
- The ease for marketers to include derived datasets in their audiences through the power of Query Service.
Objectives objectives
This workflow demonstration relies on several ÃÛ¶¹ÊÓƵ Experience Platform services. If you want to follow along, it is recommended to have a good understanding of the following features and services:
The browse abandonment example centers on using ÃÛ¶¹ÊÓƵ Analytics data to create a particular actionable audience. The audience is refined to include every customer who browsed the website in the last four days but did not make a purchase. Each profile in the audience is then targeted with the highest-price SKU that resulted from the customer’s behavior pattern.
The query itself is very prescriptive and only includes data that meets the use case criteria for the segment definition. This improves performance by minimizing the amount of Analytics data being processed. It also orders the data by price from highest to lowest and chooses the highest-priced SKU that the user was browsing.
The query used in the presentation can be seen below:
INSERT INTO summit_adv_data_prep_dataset
SELECT STRUCT(
customerId AS crmCustomerId, struct(sku AS sku, price AS sku_price, abandonTS AS abandonTS) AS abandonBrowse) AS _pfreportingonprod
FROM
(SELECT
B.personKey.sourceId,
A.productListItems[0].SKU AS sku,
max(A.timestamp) AS abandonTS,
max(c._pfreportingonprod.price) AS price
FROM summit_adobe_analytics_dataset A,profile_attribute_14adf268_2a20_4dee_bee6_a6b0e34616a9 B,summit_product_dataset c
WHERE A._experience.analytics.customDimension.evars.evar1 = B.personKey.sourceID
AND productListItems[0].SKU = C._pfreportingonprod.sku
AND A.web.webpagedetails.URL NOT LIKE '%orderconfirmation%'
AND timestamp > current_date - interval '4 day'
GROUP BY customerId,sku
order by price desc)D;
Query Service browse abandonment example using adobe analytics video-example
The video presentation seen below provides a holistic, real-world use case for your Experience Platform data focussed on Query Service and ÃÛ¶¹ÊÓƵ analytics integrations.
My name is Alex Alinoli and I’m a senior data architect. Today I’ll be talking about advanced data processing and follow that up with a demo. To demonstrate the power of advanced data processing, we’re going to use an example called Abandoned Browse. Now an example of Abandoned Browse is imagine I had signed up for a yoga class and my yogi told me that this Luma Apparel site was the best place for new gear. So imagine I’m on the site, browsing for a couple of hours, can’t find what I like and I take a break. Me taking that break is considered an abandoned browse. I was on the site viewing products but did not ultimately purchase anything. Now a standard campaign might say, hey Alex, we saw you were on our site, why don’t you continue shopping? But with advanced data processing you could say something like, Hey Alex, how about you crush that first yoga session with these buttery soft lime green yoga pants? If I saw that, I’d be like, wow guys, you totally get me. Now you might be thinking, can a database query really ignite that kind of passion? The answer is yes. Because think about what went into generating that message. We use advanced data processing to analyze this person’s web behavior, to understand what products they’re viewing and calculate the most expensive thing they looked at were lime green yoga pants. And we use those data points to provide a personalized and targeted message. Now we’ve talked about how advanced data processing can help your marketing campaigns. Let’s go and talk to you how we can build this audience in AEP. The first thing we need to understand is the data architecture. The data architecture refers to understanding what tables you have available to you and which ones you may need to build your audience and how they’re all connected together. So in our case, we need three key pieces of data. We need our browsing history, which would come from our analytics data. We need our product data, which will give us the price of the SKUs that our customers have been browsing. And we need our customer data. And customer data can come from an out of the box table called profile attributes, which is our unified view of the profile. So now that we know what tables we need to build on our query, before we actually go and build out the query, we need to understand the quality of the data that we’ll be using. Right? So the analytics data is really the key here. It is what’s joining our profile to our product, to our web browsing behavior. So let me drill into the analytics schema, and we can take a look at how data quality is important. First I’ll talk about is this SKU field. This is what we’re using to join our analytics data to product. And you may think because it’s called SKU that it always has SKU. But you’d be wrong. Right? So we’ve seen this populated with device IDs or product IDs or any other non-SKU data. Just because it’s called SKU, or just as any field is called whatever it’s called, you cannot assume that the data contained within is quality data. Thankfully, data processing gives us some tools to mitigate these kind of issues. For example, if we knew we had different types of data coming into SKU, but we knew that valid SKU data was, let’s say, alphanumeric, in data processing, in our query, we could enforce a rule saying, I only want to capture data that has an alphanumeric pattern. That way I maintain the quality of my query, which would result in a quality audience. Similarly, if I knew I had device and product data that I didn’t want to include in my query, I could always write a criteria to exclude that information as well. So you always want to make sure we know what data is flowing into our datasets so we know that when we generate our audience, how reliable it’s going to be. So talk about architecture, the data quality, and before we get into the actual demo, I want to talk about why we don’t even need data processing. We talked about earlier that we can use it to compute attributes, which allows us to have a very dynamic way of segmenting and personalizing our messages. But outside of that, we don’t have any ability in even segmentation destinations to calculate these data points on the fly. That’s why it’s important that we need data processing. Similarly, if we didn’t have the feature of data processing, what would you have to do to get this kind of personalized information? You would have to take data out of AEP, manipulate it, massage it, and then bring it back in. And if you’re talking about a couple of attributes here and there, maybe that’s not a big deal. But if you’re talking about hundreds of attributes or hundreds of campaigns, that can greatly impact your project timeline. You have to ship data in and out constantly every time you want to build something new. And that goes for the marketer as well. If they’re in the platform and they know you have the data available, they don’t want to have to worry about having to contact other teams to aggregate this data so they can build out their campaigns. So AEP provides you both the data repository and the tools you need to activate upon your data in the most efficient and reliable way possible. So having said that, let’s talk about how we could actually build this audience in the tool. And the first thing we’ll talk about is the actual query you want to build out. Let me open that up. So in AEP, there are two ways to execute a query. One is via the UI, and the other is via a command line prompt. I prefer command line prompts, so that’s what I’ll be demoing today. Let’s talk about the actual query that we’re looking to build out. This one has two main sections to it. We have a select statement, which is generating our abandoned browse audience. And then we have an insert statement, which is saving this audience into a custom table. Let’s first focus on our select statement. So you can see here, we’re collecting four key data points, customer ID, SKU, the time they abandoned, and the price of the SKU. We’re then connecting the three tables we talked about earlier, analytics, unified profile, and product. And we’re then connecting analytics to profile, and then connecting analytics to product. And then you’ll see here, we have two more conditions. This is an abandoned browse use case, so we want to avoid including any data containing order confirmation pages, because that would imply that they’ve actually gone ahead and purchased an actual IDLE. Next part on timestamp is analytics data is usually pretty huge. We always want to find ways to minimize the amount of data that we’re crunching for performance reasons and just for a use case. We don’t need people who abandoned browse like 20 days ago, 15 days ago. We want to be very prescriptive in these queries. In this case, we’ve chosen an interval of four days. So it gives us the flexibility in our segmentation. We can choose anyone who’s abandoned browse in the past hour, two hours, two days, up to four days ago. This is a flexible rolling four-day window in which to choose our audiences. And then you’ll see at the end, we have an ordering. So this ties back to our original example where we’re capturing the most expensive SKU they browsed. So we’re ordering our data by price from highest to lowest and choosing the highest priced SKU they were browsing. At the top, we have our insert statement. This is essentially saving our audience into a table called Summit Advanced Data Prep Dataset. This shows you that we can create schemas and tables not only for incoming data feeds or for streaming data, we can actually create a query that calculates computed attributes and save those attributes in its own custom table. And we can take that table, enable it for profile, and then use these data points, customer ID or SKU or abandoned timestamp or SKU price, use those attributes for segmentation and in personalization. Having said that, let’s talk about how we can actually execute this query. So as I mentioned, I prefer the command line interface, and that’s what I’ll show you in today’s demo. So to execute it, we can go to the query screen and we go to the credentials tab. And this gives you the connection string you need to connect to the database. Now, if anyone’s a little worried about giving people access to connect to the production database, you can absolutely restrict access to this page. Okay, so I’m going to copy my connection string. I will open up a command prompt and then simply paste it. And now I’m obviously my token, so we can refresh our page and get a new token generated. And I will copy this and I’ll open this up and put this in there. Enter. And I’m now properly connected to the database. So now that I’m in, I can just simply copy paste my SQL statement and execute it. So I’ll go back to our query. Copy it. Put it in here. And hit enter. Now we can execute this query directly in this window and see the results. Now, queries are not always run in a bubble. In our case, let’s say we want to schedule this to run every day. We can easily do that in the UI as well. I’ll go to under query still. I go under browse. I can click on create query. Enter my query that I want to schedule. Give it a name. Let’s say Summit data processing. It’s safe. Now I have my query saved. Now that it’s saved, I can click into it and I now have the ability to add a schedule to it. We talked about creating a daily. You have the ability to choose your frequency, start and end dates, and the data set where you want to save the output of your query. So in here, I’m going to choose a frequency of daily. I’m going to choose every one day. And maybe I’ll have this run until the end of the year. And then I want to save the output to my advanced data prep schema. And I click save. And save again. And I’m all set. So it’s that easy to create and execute a query in the command prompt. And I also scheduled to run on your desired frequency. Daily, hourly, weekly, etc. Let’s take a look at the resulting schema that we just talked about. We’re saving everything into a schema called advanced data prep. You’ll see in this schema, it matches the data points that we are querying in our SQL statement. The output of the SQL query will populate the data set with these four data points. And we can now use these in our segmentation and destinations. You’ll see also this is profile enabled, which is key. So we have our query. It’s scheduled. We’ve built our schema to accept the output of that query. Now let’s see how we can actually use that audience in a segment. Let’s go to segments. Band and browse. And let’s take a look at the criteria. So in our segment window, there’s two main panels. There’s the profile attribute panel. And then there’s the event panel at the bottom. Let’s first start at the profile panel. You can see here, we’re looking to only select customers who have abandoned in the past four days. You can see this abandoned timestamp is the attribute that we calculated in our query. You can also see under attributes that go under my tenant. You’ll see my abandoned browse attributes are right here. So you can see I can just use these directly in my segmentation criteria for any sort of logic I want to perform. In this case, I want to use the abandoned timestamp, make sure that the people I’m selecting have only abandoned in the past four days. And on top of that, I want to also exclude anybody that’s hit the order confirmation page in the past four days as well. So I have a clear cut audience of people who have abandoned browse without having made a purchase. Now you may be looking at this and be like, this looks super simple. This can’t be real. But it is because that’s the beauty of data processing. Having the query, the query is doing all the hard work behind the scenes. So you can just focus on the data points you care about. Right. Some customers may have, you know, five, six, seven attributes that they always need to include on every segment, like emailable yes, no, or certain preferences. And there could be a number of static criteria that are always needed. One of the other advantages of data processing is that you’re not limited just using it to compute attributes. You can include logic in your query that satisfies some of these data conditions you always have. You can include the email constraints or preference constraints in your query itself. So all that work is done on the back end. And so it greatly simplifies how you build out your segments. And so the last part is we’re going to talk about how we can actually activate this data. We’ve now segmented. And that is via destinations.
So you can see here I have my segment that activated it to an S3 location. So here we’ll look at what the attributes are for personalization. Right. So I have my destination configured. I have selected my abandoned browse segment. And now this is asking me what attributes do I want to send in the data feed that I create on the S3 location. And you can see here I am selecting all the computer attributes that I generated from my SQL query. Picking the customer ID, the SKU price, the SKU, and the abandoned timestamp. These can all now be used in personalization downstream. For example, if an email service provider can use these attributes to populate an email. If there was more metadata I want to include here, such as the category or the name of the item, I could simply add that to my query as data points and then use them here to export them into destination. So your data processing logic has many functions. You can use it for accommodating complex logic and for segmentation. You can use it for calculating various personalized attributes we use downstream. And using it to greatly simplify how you build out your segments. All right. So that brings us to the end of this demo. We’ve talked through why data processing is key. An example of having used it for an abandoned browse situation. And that how we want to build that query depends on our existing data architecture. Ensuring that the quality of the data meets our needs. And if it doesn’t, how we can mitigate that. And how we can actually schedule this query to run on a certain frequency. And then use it downstream in segmentation and destinations for personalization. So thank you all for your time today. I hope this was informative.
Benefits of Query Service benefits
The features provided by Query Service serves many purposes. You can use it to accommodate complex logic for segmentation, for calculating various personalized attributes for use downstream, or to greatly simplify how you build out your audiences.
Query Service enables you to include constraints in your queries to simplify your audience building process. This improves data quality by ensuring the right data qualifies for your audiences. Maintaining the quality of your query results in an accurate audience and helps with data reliability. You can also save your audience by creating schemas and custom tables based on attributes derived from your query. A custom table can be enabled for Profile and you can use these data points for segmentation and personalization. This feature assists marketers who want to create a clear-cut audience of people.
Also, by including logic in your query that satisfies any recurring or static conditions, Query Service extracts the burden of elaborate segmentation.
ÃÛ¶¹ÊÓƵ Experience Platform provides a data repository and the necessary tools to activate your data in an efficient and reliable way. By keeping data inside Platform, it allows you to derive attributes while running other processes and removes the need to export data to third-party tools for manipulation and processing. Such processing overheads can greatly impact a project timeline when dealing with hundreds of attributes or campaigns. This gives marketers a single location to access their data and build out campaigns as well as a very dynamic means of segmenting and personalizing their messages.
Next steps
By reading this document, you should now understand how Query Service impacts not only the quality of your data and ease of segmentation but also its importance within your data architecture for the entire end-to-end workflow. For more applicable SQL examples that use ÃÛ¶¹ÊÓƵ Analytics with Query Service, see the ÃÛ¶¹ÊÓƵ Analytics merchandising variables use case.
Other documents that demonstrate the benefits of Query Service to your organization’s strategic business insights are the bot filtering use case example.
Alternatively, these documents can benefit your understanding of Query Service features: