Streams microphone data from a PyAudio device (Documentation)
Starts an arecord process locally and reads audio data from its standard out (Documentation)
Calls an external program to record raw audio (Documentation)
Expects audio frames from a custom external service that supports the Hermes Protocol (Documentation)
Note: statistics will not work if UDP audio is enabled
Listens for a keyphrase using pocketsphinx (Documentation)
Listens for one or more wake words with snowboy (Documentation)
See documentation for how to use multiple wake words.
Listens for a wake word with porcupine (Documentation)
Listens for a wake word with Mycroft Precise (Documentation)
Listens for a wake word with Raven (Documentation)
- Template Directory: profiles/it/raven
| Enabled | Keyword | Example 1 | Example 2 | Example 3 | |
|---|---|---|---|---|---|
Default Settings
Expects hotword detections from an external service that supports the Hermes Protocol (Documentation)
Does speech recognition with CMU's pocketsphinx. Less accurate, but supports many languages (Documentation)
- Acoustic Model: profiles/it/acoustic_model
- Dictionary: profiles/it/dictionary.txt
- Language Model: profiles/it/language_model.txt
- Base Dictionary: profiles/it/base_dictionary.txt
- Base Language Model: profiles/it/base_language_model.txt
- Base Language Model FST: profiles/it/base_language_model.fst
Does speech recognition with Kaldi. Fast and accurate for trained sentences (Documentation)
- Model Root: /profiles/it/kaldi/model
- Graph Directory: <model_root>/graph
- Base Dictionary: /profiles/it/kaldi/base_dictionary.txt
- Base Graph Directory: profiles/it/base_graph
- Base Language Model: profiles/it/kaldi/base_language_model.txt
- Base Language Model FST: profiles/it/base_language_model.fst
Does speech recognition with Mozilla's DeepSpeech version 0.9. Slower on some hardware, but often more accurate (Documentation)
- Acoustic Model: profiles/it/deepspeech/model/0.9/output_graph.pbmm
- Language Model: profiles/it/deepspeech/lm.binary
- Scorer: profiles/it/deepspeech/scorer
- Base Language Model: profiles/it/deepspeech/model/0.9/base_lm.binary
- Base Scorer: profiles/it/deepspeech/model/0.9/base.scorer
- Base Language Model FST: profiles/it/base_language_model.fst
Does speech recognition with Vosk. Fast and accurate for open transcription (Documentation)
- Model Root: profiles/it/vosk/model
- Words JSON: profiles/it/vosk/words.json
POSTs WAV audio to a remote HTTP endpoint, expecting a plain text transcription back (Documentation)
Calls an external program with WAV audio on standard input and expects a text transcription on standard output (Documentation)
Expects an external service that supports the Hermes Protocol to do speech to text (Documentation)
Voice Command Settings
Uses rhasspy-nlu to recognize only the sentences Rhasspy was trained on (Documentation)
- Intent Graph: /profiles/it/intent_graph.pickle.gz
- Stop Words: /usr/lib/rhasspy/rhasspy-profile/rhasspyprofile/profiles/it/stop_words.txt
Finds the closest matching intent using rapidfuzz library (Documentation)
Note: requires an external service (installation)
- Examples File: profiles/it/intent_examples.md
- Intent Graph: /profiles/it/intent_graph.pickle.gz
Uses Snips NLU to flexibly recognize sentences (Documentation)
- Engine: profiles/it/snips/engine
- Dataset: profiles/it/snips/dataset.yaml
POSTs plain text to an HTTP endpoint and receives intent JSON back (Documentation)
Calls an external program with text on standard input and expects intent JSON on standard output (Documentation)
Expects an external service that supports the Hermes Protocol to recognize intents (Documentation)
Uses eSpeak to speak sentences. Sounds robotic, but supports many languages and locales (Documentation).
Uses a remote MaryTTS web server to speak sentences (Documentation)
Note: requires an external service (Docker image available)
Note: requires an internet connection and a Google account
- Credentials: profiles/it/tts/googlewavenet/credentials.json
- WAV cache: profiles/it/tts/googlewavenet/cache
Uses a remote OpenTTS web server to speak sentences (Documentation)
Note: requires an external service (Docker image available)
POSTs text to a remote HTTP endpoint and plays received WAV audio (Documentation)
Calls an external program with text on standard input and plays WAV audio from standard output (Documentation)
Expects an external service that supports the Hermes Protocol to speak sentences (Documentation)
POSTs WAV audio to a remote HTTP endpoint (Documentation)
Expects an external service that supports the Hermes Protocol to play audio (Documentation)
Makes Rhasspy behave like a typical voice assistant, automatically listening for voice commands after hotword detections and recognizing intents from them (Documentation)
Uses rhasspy-dialogue-hermes service
Expects a custom external service that supports the Hermes Protocol to do something with hotword detections, etc. (Documentation)
Sends intents or events directly to Home Assistant or Hass.io (Documentation)
POSTs intent JSON to a remote HTTP endpoint (Documentation)
Sounds
WAV files to play when Rhasspy wakes up and is finished recording a voice command.
Note: dialogue management must be set to "Rhasspy" to hear these sounds
Use ${RHASSPY_PROFILE_DIR} environment variable for your profile directory.
Certificates
Files needed for using HTTPS with your Rhasspy server or Home Assistant
Note: all paths should be absolute
Use ${RHASSPY_PROFILE_DIR} environment variable for your profile directory.