Is it possible to have a breakdown of which utterances are used the most?
I'd like to save that information in order to further extend and improve the list of utterances mapped to each intent.
From Amazon Alexa Measure dashboard I can see a breakdown of intents but not utterances.
Is this a limitation of the dashboard or is the data not returned by the API at all?
Unfortunately Alexa does not give us information on what utterances the user asked. It only maps the utterances to the specific intent.
One possible solution would be to create a custom slot and a slot type and have all the utterance in the slot type and have an utterance like {slot} in the intent.
This way the utterance will be sent in the slot value to your back-end API which you can use for measuring.
Related
Is it possible for alexa user to have different responses based on config in the app. For example my skill is returning measurements. Some users may prefer metric and others imperial. I'd like users to be able to specify this (and may be some other things) to give a personalised experience. Can this be configured in the Amazon Alexa app?
I was thinking I might have to have some persistent storage for this (DDB for example) which would mean the app would write to the DDB and the skill would read from it to get the personalised response.
Thanks
Can this be configured in the Amazon Alexa app?
Unfortunately not in the way which you seem to be suggesting.
If you really wanted users to set preferences through the app, this could be done through account linking. However, it is generally discouraged (Alexa is meant to be "Voice-First") and likely to present additional obstacles if what you're wanting to do is allow users to set preferences for different devices.
However, using persistent storage for user preferences in generally is a good idea and as you've suggested, DynamoDB can do this.
If you take this approach you could ask users what their preferences are the first time they use a skill on a device and store this together with the device ID.
There is some good information about device ID in the Amazon documentation and some helpful tips here:
Get unique device id for every amazon echo devices
I have read many articles which specifies flash briefing as a separate skill and custom skill as an independent skill
Is it possible to combine flash briefing and custom skill together or is it not available right now?
Currently, Alexa don't have a way to do it in a single skill per:
https://forums.developer.amazon.com/questions/159904/alexa-can-we-combine-flash-briefing-and-conversati.html
My understanding is that Amazon ASK still does not provide:
The raw user input
An option for a fallback intent
An API to
dynamically add possible options from which Alexa can be better
informed to select an intent.
Is this right or am I missing out on knowing about some critical capabilities?
Actions on Google w/ Dialogflow provides:
raw user input for analysis: request.body.result.resolvedQuery
fallback intents:
https://dialogflow.com/docs/intents#fallback_intents
An APi to dynamically add user expressions (aka sample utterances): PUT
/intents/{id}
These tools provide devs with the ability to check to see if the identified intent is correct and if not fix it.
I know there have been a lot of questions asked previously, just a few here:
How to add slot values dynamically to alexa skill
Can Alexa skill handler receive full user input?
Amazon Alexa dynamic variables for intent
I have far more users on my Alexa skill than my AoG app simply because of Amazon's dominance to date in the market - but their experience falls short of a Google Assistant user experience because of these limitations. I've been waiting for almost a year for new Alexa capabilities here, thinking that after Amazon's guidance to not use AMAZON.LITERAL there would be improvements coming to custom slots. To date it still looks like this old blog post is still the only guidance given. With Google, I dynamically pull in utterance options from a db that are custom for a given user following account linking. By having the user's raw input, I can correct the choice of intent if necessary.
If you've wanted these capabilities but have had to move forward without them, what tricks do you have to get accurate intent handling with Amazon when you don't know what the user will say?
EDIT 11/21/17:
In September Amazon announced the Alexa Skill Management API (SMAPI) which does provide the 3rd bullet above.
Actually this should be better a comment but i write to less at stackoverflow to be able to comment. I am with you on all.
But Amazons Alexa has also a very big advance.
The intent Schema is seeming to directly influence the Voice to Text recognition. Btw. can someone confirm if this is correct?
At Google Home it seems not to be the case.
So matching of unusual names is even more complicated than at alexa.
And it sometimes just recognize absolute bullshit.
Not sure which I prefer currently.
My feeling is for small apps is Alexa much better, because it better match the Intent phrases when it has lesser choices.
But for large Intent schemas, it get really trouble and in my tests some of the intents were not matched at all correct.
Here the google home and action SDK wins, probably? Cause Speech to text seem to be done before and than a string pattern to intent schema matching is happening. So this is probably more robust for larger schemas?
To get something like an answer on your questions:
You can try to add as much as possible that can be said to a slot. And than match the result from the Alexa request to your database via Jaro winkler or some other string distance.
Was I tried for Alexa was to find phrases that are close to what the user say. And this i added as phrases to fill a slot.
So a module in our webpage was an intent in the schema. And Than I requested To say what exactly should be done in that module (this was the slot filling request). The Answer was the slot filling utterance.
For me that was slightly better working than the regulary intent schema. But it require more talking so i dont like it so much.
Let me go straight to answering your 3 questions:
1) Alexa does provide the raw input via the slot type AMAZON.Literal but it's now deprecated and you're advised to use AMAZON.SearchQuery for free form capture. However, if instead of using SearchQuery you define a custom slot type and provide samples (training data) the ASR will work better.
2) Alexa supports FallbackIntent since I believe May 2018. The way it work is by automatically generating a model for your skill where out-of-domain requests are routed through a fallback intent. It works well
3) Dynamically adding slot type values is not feasible since when you provide samples you're really providing training data for a model than will be able to then process similar values beyond the ones you defined. If you noticed when you provide a voice interaction model schema then you have to build the model (in this step the training data provided in the samples is used to create the model). One example, when you define a custom slot of type "Car" and you provide the samples "Toyota", "Jeep", "Chevrolet" and "Honda" then, the system will also go to the same intent if the user says "Ford"
Note: SMAPI does allow to get and update the interaction model, so technically you could download the model via API, modify it with new training data, upload it again and rebuild the model. This is kind of awkward though
Sometimes in my Skill, after receiving an intent, Alexa will ask back to clarify. For example: "How many do you want to add?" In this case, the answer of the user will likely match only one Intent, (or the StopIntent).
Can I use this knowledge to improve the recognition of the answer?
At this time, there is no way to dynamically alter the list of intents.
In your application, however, you should check to see if you received a valid intent for the current context and provide a suitable error message. I have heard of some people looking at easily confused intents, and making a guess that a similar, contextually correct, intent was chosen.
Is there a way to block compete.com or alexa from getting the traffic on my website? For competitive reasons, I don't want anyone knowing that info. I don't have advertising, just a paid-only site....would blocking compete.com or alexa in anyway affect my search engine rankings?
Is there a way to block compete.com or alexa from getting the traffic on my website
As far as I know, Alexa and Compete.com both use user-installed toolbars that analyze what sites the user visits (similar to how TV Ratings are done; see e.g. the Wikipedia article on the Alexa Toolbar). They do not fetch your traffic numbers as such; they estimate their own.
To mess with their counting process, you would have to block users with the Alexa Toolbar installed. While that may be possible, it's foolish in the extreme: You would be locking out innocent users, and decrease your Alexa ratings, which wouldn't affect them in the slightest.
Yes you can do it easily and automatically without having to contact them.
For Alexa, you can do it using robot.txt
User-agent: ia_archiver
Disallow: /
The toolbar plugin is used to monitor traffic.
I don't have details for the others, but I've just found some information about robot.txt for Compete, Quantcast and Alexa: look a the end of this article for practical details.
Please note that by adding that in your robot.txt it will also force them to remove all data they gathered retroactively.
This is useful for websites like archive.org (linked to Alexa, in fact ia in ia_archiver stands for Internet Archive). They explain the process here: http://www.archive.org/about/exclude.php
Please note that in most countries, being able to remove your data from databases is a legal right.