Quickstart: Run the Speech Devices SDK sample app on Linux

In this quickstart, you'll learn how to use the Speech Devices SDK for Linux to build a speech-enabled product or use it as a Conversation Transcription device. Currently only the Azure Kinect DK is supported.

The application is built with the Speech SDK package, and the Eclipse Java IDE (v4) on 64-bit Linux (Ubuntu 16.04, Ubuntu 18.04, Debian 9). It runs on a 64-bit Java 8 runtime environment (JRE).

This guide requires an Azure Cognitive Services account with a Speech service resource. If you don't have an account, you can use the free trial to get a subscription key.

The source code for the sample application is included with the Speech Devices SDK. It's also available on GitHub.


This quickstart requires:

  • Operating System: 64-bit Linux (Ubuntu 16.04, Ubuntu 18.04, Debian 9)
  • Azure Kinect DK
  • Eclipse Java IDE
  • Java 8 or JDK 8 only.
  • An Azure subscription key for the Speech service. Get one for free.
  • Download the latest version of the Speech Devices SDK for Java, and extract the .zip to your working directory.


    The JRE-Sample-Release.zip file includes the JRE sample app and this quickstart assumes that the app is extracted to /home/wcaltest/JRE-Sample-Release

Make sure these dependencies are installed before starting Eclipse.

  • On Ubuntu:

    sudo apt-get update
    sudo apt-get install libssl1.0.0 libasound2
  • On Debian 9:

    sudo apt-get update
    sudo apt-get install libssl1.0.2 libasound2

Conversation Transcription is currently only available for "en-US" and "zh-CN", in the “centralus” and “eastasia” regions. You must have a speech key in one of those regions to use Conversation Transcription.

If you plan to use the intents you'll need a Language Understanding Service (LUIS) subscription. To learn more about LUIS and intent recognition, see Recognize speech intents with LUIS, C#. A sample LUIS model is available for this app.

Create and configure the project

  1. Start Eclipse.

  2. In the Eclipse IDE Launcher, in the Workspace field, enter the name of a new workspace directory. Then select Launch.

    Screenshot of Eclipse Launcher

  3. In a moment, the main window of the Eclipse IDE appears. Close the Welcome screen if one is present.

  4. From the Eclipse menu bar, create a new project by choosing File > New > Java Project. If not available choose Project and then Java Project.

  5. The New Java Project wizard starts. Browse for the location of the sample project. Select Finish.

    Screenshot of New Java Project wizard

  6. In the Package explorer, right-click your project. Choose Configure > Convert to Maven Project from the context menu. Select Finish.

    Screenshot of Package explorer

  7. Open the pom.xml file and edit it.

    At the end of the file, before the closing tag </project>, create repositories and dependencies elements, as shown here, and ensure the version matches your current version:

             <name>Microsoft Cognitive Services Speech Maven Repository</name>
  8. In the Package explorer, right-click your project. Choose Properties, then Run/Debug Settings > New… > Java Application.

  9. The Edit Configuration window appears. In the Name field enter Main, and use Search for the Main Class to find and select com.microsoft.cognitiveservices.speech.samples.FunctionsList.

    Screenshot of Edit Launch Configuration

  10. Copy the audio binaries for your target architecture, from either Linux-arm or Linux-x64, to the Java Project location, eg /home/wcaltest/JRE-Sample-Release

  11. Also from the Edit Configuration window select the Environment page and New. The New Environment Variable window appears. In the Name field enter LD_LIBRARY_PATH and in the value field enter the folder containing the *.so files, for example /home/wcaltest/JRE-Sample-Release

  12. Copy kws.table and participants.properties into the project folder target/classes

Configure the sample application

  1. Add your speech subscription key to the source code. If you want to try intent recognition, also add your Language Understanding service subscription key and application ID.

    For speech and LUIS, your information goes into FunctionsList.java:

     // Subscription
     private static String SpeechSubscriptionKey = "<enter your subscription info here>";
     private static String SpeechRegion = "westus"; // You can change this if your speech region is different.
     private static String LuisSubscriptionKey = "<enter your subscription info here>";
     private static String LuisRegion = "westus2"; // you can change this, if you want to test the intent, and your LUIS region is different.
     private static String LuisAppId = "<enter your LUIS AppId>";

    If you are using conversation transcription, your speech key and region information are also needed in Cts.java:

     private static final String CTSKey = "<Conversation Transcription Service Key>";
     private static final String CTSRegion="<Conversation Transcription Service Region>";// Region may be "centralus" or "eastasia"
  2. The default keyword (keyword) is "Computer". You can also try one of the other provided keywords, like "Machine" or "Assistant". The resource files for these alternate keywords are in the Speech Devices SDK, in the keyword folder. For example, /home/wcaltest/JRE-Sample-Release/keyword/Computer contains the files used for the keyword "Computer".


    You can also create a custom keyword.

    To use a new keyword, update the following line in FunctionsList.java, and copy the keyword to your app. For example, to use the keyword 'Machine' from the keyword package machine.zip:

    • Copy the kws.table file from the zip package into the project folder target/classes.

    • Update the FunctionsList.java with the keyword name:

      private static final String Keyword = "Machine";

Run the sample application from Eclipse

  1. From the Eclipse menu bar, Run > Run

  2. The Speech Devices SDK example application starts and displays the following options:

    Sample Speech Devices SDK example application and options

  3. Try the new Conversation Transcription demo. Start transcribing with Session > Start. By default everyone is a guest. However, if you have participant’s voice signatures they can be put into participants.properties in the project folder target/classes. To generate the voice signature, look at Transcribe conversations (SDK).

    Demo Conversation Transcription application

Create and run standalone the application

  1. In the Package explorer, right-click your project. Choose Export.

  2. The Export window appears. Expand Java and select Runnable JAR file and then select Next.

    Screenshot of the Export window

  3. The Runnable JAR File Export window appears. Choose an Export destination for the application, and then select Finish.

    Screenshot of Runnable JAR File Export

  4. Please put kws.table and participants.properties in the destination folder chosen above as these files are needed by the application.

  5. Set the LD_LIBRARY_LIB to the folder containing the *.so files

    export LD_LIBRARY_PATH=/home/wcaltest/JRE-Sample-Release
  6. To run the standalone application

    java -jar SpeechDemo.jar

Next steps