Quickstart: Run the Speech Devices SDK sample app on Android

In this quickstart, you'll learn how to use the Speech Devices SDK for Android to build a speech-enabled product or use it as a Conversation Transcription device.

This guide requires an Azure Cognitive Services account with a Speech service resource. If you don't have an account, you can use the free trial to get a subscription key.

The source code for the sample application is included with the Speech Devices SDK. It's also available on GitHub.


Before you start using the Speech Devices SDK, you'll need to:

Set up the device

  1. Start Vysor on your computer.


  2. Your device should be listed under Choose a device. Select the View button next to the device.

  3. Connect to your wireless network by selecting the folder icon, and then select Settings > WLAN.

    Vysor WLAN


    If your company has policies about connecting devices to its Wi-Fi system, you need to obtain the MAC address and contact your IT department about how to connect it to your company's Wi-Fi.

    To find the MAC address of the dev kit, select the file folder icon on the desktop of the dev kit.

    Vysor file folder

    Select Settings. Search for "mac address", and then select Mac address > Advanced WLAN. Write down the MAC address that appears near the bottom of the dialog box.

    Vysor MAC address

    Some companies might have a time limit on how long a device can be connected to their Wi-Fi system. You might need to extend the dev kit's registration with your Wi-Fi system after a specific number of days.

Run the sample application

To validate your development kit setup, build and install the sample application:

  1. Start Android Studio.

  2. Select Open an existing Android Studio project.

    Android Studio - Open an existing project

  3. Go to C:\SDSDK\Android-Sample-Release\example. Select OK to open the example project.

  4. Add your speech subscription key to the source code. If you want to try intent recognition, also add your Language Understanding service subscription key and application ID.

    For speech and LUIS, your information goes into MainActivity.java:

     // Subscription
     private static String SpeechSubscriptionKey = "<enter your subscription info here>";
     private static String SpeechRegion = "westus"; // You can change this if your speech region is different.
     private static String LuisSubscriptionKey = "<enter your subscription info here>";
     private static String LuisRegion = "westus2"; // you can change this, if you want to test the intent, and your LUIS region is different.
     private static String LuisAppId = "<enter your LUIS AppId>";

    If you are using conversation transcription, your speech key and region information are also needed in conversation.java:

     private static final String CTSKey = "<Conversation Transcription Service Key>";
     private static final String CTSRegion="<Conversation Transcription Service Region>";// Region may be "centralus" or "eastasia"
  5. The default keyword is "Computer". You can also try one of the other provided keywords, like "Machine" or "Assistant". The resource files for these alternate keywords are in the Speech Devices SDK, in the keyword folder. For example, C:\SDSDK\Android-Sample-Release\keyword\Computer contains the files used for the keyword "Computer".


    You can also create a custom keyword.

    To use a new keyword, update the following two lines in MainActivity.java, and copy the keyword package to your app. For example, to use the keyword 'Machine' from the keyword package kws-machine.zip:

    • Copy the keyword package into the folder “C:\SDSDK\Android-Sample-Release\example\app\src\main\assets\”.

    • Update the MainActivity.java with the keyword and the package name:

      private static final String Keyword = "Machine";
      private static final String KeywordModel = "kws-machine.zip" // set your own keyword package name.
  6. Update the following lines, which contain the microphone array geometry settings:

    private static final String DeviceGeometry = "Circular6+1";
    private static final String SelectedGeometry = "Circular6+1";

    This table lists supported values:

    Variable Meaning Available values
    DeviceGeometry Physical mic configuration For a circular dev kit: Circular6+1
    For a linear dev kit: Linear4
    SelectedGeometry Software mic configuration For a circular dev kit that uses all mics: Circular6+1
    For a circular dev kit that uses four mics: Circular3+1
    For a linear dev kit that uses all mics: Linear4
    For a linear dev kit that uses two mics: Linear2
  7. To build the application, on the Run menu, select Run 'app'. The Select Deployment Target dialog box appears.

  8. Select your device, and then select OK to deploy the application to the device.

    Select Deployment Target dialog box

  9. The Speech Devices SDK example application starts and displays the following options:

    Sample Speech Devices SDK example application and options

  10. Try the new Conversation Transcription demo. Start transcribing with 'Start Session'. By default everyone is a guest. However, if you have participant’s voice signatures they can be put into a file /video/participants.properties on the device. To generate the voice signature, look at Transcribe conversations (SDK).

    Demo Conversation Transcription application

  11. Experiment!


If you cannot connect to the Speech Device. Type the following command in a Command Prompt window. It will return a list of devices:

 adb devices


This command uses the Android Debug Bridge, adb.exe, which is part of the Android Studio installation. This tool is located in C:\Users[user name]\AppData\Local\Android\Sdk\platform-tools. You can add this directory to your path to make it more convenient to invoke adb. Otherwise, you must specify the full path to your installation of adb.exe in every command that invokes adb.

If you see an error no devices/emulators found then check your USB cable is connected and ensure a high quality cable is used.

Next steps