Name-free interaction enables customers to interact with Alexa without invoking a specific skill by name, which helps facilitate greater interaction with Alexa because customers do not always know which skill is appropriate.

When Alexa receives a request from a customer without a skill name, such as “Alexa, play relaxing sounds with crickets,” Alexa looks for skills that might fulfill the request. Alexa determines the best choice among eligible skills and hands the request to the skill.

To make your skill more discoverable for name-free interaction, you can implement the CanFulfillIntentRequest interface in your skill.

In Voxa, you can take advantage of this feature by following this example:

skill.onCanFulfillIntentRequest((alexaEvent, reply) => {
  if ( === 'InfluencerIntent') {

    _.each(alexaEvent.intent.params, (value, slotName) => {
      reply.fulfillSlot(slotName, 'YES', 'YES');

  return reply;

Voxa offers the function onCanFulfillIntentRequest so you can implement it in your code to validate wether you’re going to fulfill the request or not.

Additionally, if you have several intents that you want to automatically fulfill, regardless of the slot values in the request, you can simply add an array of intents to the property: defaultFulfillIntents of the Voxa config file:

const defaulFulfillIntents = [

const skill = new StateMachineSkill({ variables, views, defaultFulfillIntents });

If Alexa sends an intent that you didn’t register with this function, then you should implement the onCanFulfillIntentRequest method to handle it. Important: If you implement this method in your skill, you should always return the reply object.

If a skill has implemented canFulfillIntent according to the interface specification, the skill should be aware that the skill is not yet being asked to take action on behalf of the customer, and should not modify any state outside its scope, or have any observable interaction with its own calling functions or the outside world besides returning a value. Thus, the skill should not, at this point, perform any actions such as playing sound, turning lights on or off, providing feedback to the customer, committing a transaction, or making a state change.

Testing using ASK-CLI

There are 2 options to test this feature manually. The first one is using the Manual JSON section of the Test tab in the developer portal. And the other one, is to use the ASK CLI from Amazon.

You can just trigger this command in the console, and you’ll get the result in your terminal:

$ ask api invoke-skill --skill-id amzn1.ask.skill.[unique-value-here] --file /path/to/input/json --endpoint-region [endpoint-region-here]