ALTextToSpeech

NAOqi Audio - Overview | API | Tutorial


What it does

The ALTextToSpeech module allows the robot to speak. It sends commands to a text-to-speech engine, and authorizes also voice customization. The result of the synthesis is sent to the robot’s loudspeakers.

How it works

Speech engines

ALTextToSpeech is based on speech synthesizers - or speech engines.

According to the selected language, a specific engine is used:

  Uses ...
Japanese microAITalk engine, provided by AI, Inc.
Other languages an engine provided by ACAPELA.

Customization

Using parameters

The output audio stream can be modified.

For example, these effects are available:

  • pitch shifting effect modifies the initial pitch of the voice,
  • double voice effect produces a “delay/echo” effect on the main voice.

additional parameters are available for microAITalk engine.

Further information can be found here : ALTextToSpeechProxy::setParameter()

Using tags - ACAPELA only

The ACAPELA’s voices can be customized through “tags” included within the string command sent to the engine. Further informations can be found here: Acapela Mobility Text TAGS Documentation

Getting Started

The easiest way to get started with ALTextToSpeech is to use the Say Choregraphe box.

Testing on a real or a virtual robot

ACAPELA and microAITalk engines are only available on the real robot.

When using a virtual robot, said text can be visualized in Choregraphe Robot view and Dialog panel.