Skip to content

shannonlowder.com

Menu
  • About
  • Biml Interrogator Demo
  • Latest Posts
Menu

Alexa Telemetry Data

Posted on February 8, 2018November 14, 2022 by slowder

When you build an Alexa skill you’ve got two rich sources of information describing how users are interacting with your skill.  The first is the data in the Alexa Skills Kit request object.  The request object is the JSON representation of what you said to your Alexa device. Alexa Voice Service(AVS) takes your spoken words and breaks them down into this object based on the interaction model you define.  This JSON is what’s sent from AVS to your custom back-end application.  The other rich source of data is whatever you decide to collect while your application is processing requests.  Let’s explore the request object and look at some key data points stored there.

Request

  • RequestID — a unique identifier for the request coming in. If you build a skill that has several steps, then each time the user speaks to their Alexa, a new RequestID is sent for each utterance.
  • Timestamp — self-explanatory.
  • Locale — a string describing your user’s preferred language. For example, “de-DE” would indicate the user has their Alexa set to German. “en-US” is United States american, and “en-IN” would indicate India English.  This data can be useful for customizing the experience to different language demands.  For example, If a user has selected “en-GB” and they ask for the closest multi-story”, they’re asking for a parking garage.
  • Intent Name — the name of the intent AVS understood.  Usually, you will have just a handful of intents. being able to measure how popular each of them are will help you decide where you want to invest more development time.
  • Intent Slot Names and values — remember these are the variables of your skill. Recording how often different values are used may help you decide to refine your utterances.

CatchAll Intent and Slot

What if you wanted to see what users were passing as Utterances that you never expected?  You can set up an intent called “CatchAll”  this intent has a single slot and utterance: “{CatchAll}”.  In your application, you fail gracefully when a user hits this intent.  You want to tell the user, “I didn’t understand you, but I’ve sent your request on for further processing.” Then, you can log all the utterances your users send to a log file.  Every so often, you can analyze this log file and look for patterns in the data.

You may find users are using a different phrasing than you expected. You might find that users are asking followup questions you never expected.  Use this data to improve your skill and increase the likelihood users will keep using your skill!

System

  • DeviceID — a unique identifier for the physical device accessing your skill through the Alexa service.
  • Supported Interfaces — is this a dot or echo and only have audio abilities, or is this a spot and has a video screen.  Knowing how many of your users have different abilities can influence your decision to create multi-modal interfaces for your skill.

Session

  • SessionID — for each conversation a unique ID is assigned.  This ID will accompany every utterance the user sends until the skill times out or the users cancels out.  This is useful for tracking your user “funnel”. Do you have users start ordering tickets through your skill, but stop before they complete the order?  Where are they stopping? Knowing this can help you refine your interaction model making it easier for your users to complete tasks.
  • Attributes — the Swiss Army Knife of the request.  These store key-value pairs during sessions.  When you need to “remember” data between steps, this is the closest destination for that data.  If the data needs to be persisted past a session, be sure you write out the data to something a little more durable, like SQL Server!
  • User Access Token –token used to identify the current user in another system, like facebook or twitter.
  • Permissions — if the user grants you permission to see their physical address or other permissions, this is where you can find the access token you’ll use to get to that sensitive data.

Internal Application Telemetry

Once the request makes it to your application, the sky’s the limit on what you can track.  Given a unique user ID, you could track how many times that user has used your skill. n You could track what time(s) of the day he or she uses the skill.  If you built a quiz skill you could track high scores by user, region, or any other metric you’re capturing.  You could also aggregate user scores to provide questions users would find more challenging.  Anything you want to track here, you can!

Analyzing this data

Collecting the data is pretty easy, but how do we analyze it?  you’ve got two choices.  You can transform the data into a format easy for you to analyze and store that in an OLAP database.  This is fine, so long as you can handle the insert volume.  You also run the risk of painting yourself into a corner by transforming the data at first.

The alternative is to store the data raw: in text or JSON format.  Leave the transformation until later, when you know what you’re looking for.  This option is a better fit for a data lake.  Data lakes can handle higher volume inserts since they’re supported by highly distributed file systems.  The cost here, is you have to transform the data into a format for analysis on demand.  This could cost more in CPU over the long term.

Either option is valid, just be sure to store the data so you can extract the intelligence from it later!  If you’re interested in more about how to analyze Alexa data from data lakes, check out my articles on Azure Data Lake.  I’ll be growing that content over the coming year as I collect more data from my Alexa skills.

In the meantime, if you have any questions, please let me know. I’m here to help!

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Recent Posts

  • A New File Interrogator
  • Using Generative AI in Data Engineering
  • Getting started with Microsoft Fabric
  • Docker-based Spark
  • Network Infrastructure Updates

Recent Comments

  1. slowder on Data Engineering for Databricks
  2. Alex Ott on Data Engineering for Databricks

Archives

  • July 2023
  • June 2023
  • March 2023
  • February 2023
  • January 2023
  • December 2022
  • November 2022
  • October 2022
  • October 2018
  • August 2018
  • May 2018
  • February 2018
  • January 2018
  • November 2017
  • October 2017
  • September 2017
  • August 2017
  • June 2017
  • March 2017
  • February 2014
  • January 2014
  • December 2013
  • November 2013
  • October 2013
  • August 2013
  • July 2013
  • June 2013
  • February 2013
  • January 2013
  • August 2012
  • June 2012
  • May 2012
  • April 2012
  • March 2012
  • February 2012
  • January 2012
  • December 2011
  • November 2011
  • October 2011
  • September 2011
  • August 2011
  • July 2011
  • June 2011
  • May 2011
  • April 2011
  • March 2011
  • February 2011
  • January 2011
  • December 2010
  • November 2010
  • October 2010
  • September 2010
  • August 2010
  • July 2010
  • June 2010
  • May 2010
  • April 2010
  • March 2010
  • January 2010
  • December 2009
  • November 2009
  • October 2009
  • September 2009
  • August 2009
  • July 2009
  • June 2009
  • May 2009
  • April 2009
  • March 2009
  • February 2009
  • January 2009
  • December 2008
  • November 2008
  • October 2008
  • September 2008
  • August 2008
  • July 2008
  • June 2008
  • May 2008
  • April 2008
  • March 2008
  • February 2008
  • January 2008
  • November 2007
  • October 2007
  • September 2007
  • August 2007
  • July 2007
  • June 2007
  • May 2007
  • April 2007
  • March 2007
  • February 2007
  • January 2007
  • December 2006
  • November 2006
  • October 2006
  • September 2006
  • August 2006
  • July 2006
  • June 2006
  • May 2006
  • April 2006
  • March 2006
  • February 2006
  • January 2006
  • December 2005
  • November 2005
  • October 2005
  • September 2005
  • August 2005
  • July 2005
  • June 2005
  • May 2005
  • April 2005
  • March 2005
  • February 2005
  • January 2005
  • November 2004
  • September 2004
  • August 2004
  • July 2004
  • April 2004
  • March 2004
  • June 2002

Categories

  • Career Development
  • Data Engineering
  • Data Science
  • Infrastructure
  • Microsoft SQL
  • Modern Data Estate
  • Personal
  • Random Technology
  • uncategorized
© 2025 shannonlowder.com | Powered by Minimalist Blog WordPress Theme