Convert Figma logo to code with AI

snakers4 logosilero-models

Silero Models: pre-trained text-to-speech models made embarrassingly simple

5,552
349
5,552
8

Top Related Projects

14,058

Offline speech recognition API for Android, iOS, Raspberry Pi and servers with Python, Java, C# and Node

43,367

🐸💬 - a deep learning toolkit for Text-to-Speech, battle-tested in research and production

DeepSpeech is an open source embedded (offline, on-device) speech-to-text engine which can run in real time on devices ranging from a Raspberry Pi 4 to high power GPU servers.

9,572

End-to-End Speech Processing Toolkit

15,211

kaldi-asr/kaldi is the official location of the Kaldi project.

31,926

Facebook AI Research Sequence-to-Sequence Toolkit written in Python.

Quick Overview

Silero Models is an open-source project that provides pre-trained speech-to-text, text-to-speech, and voice activity detection models. It aims to make speech recognition and synthesis accessible and easy to use for developers and researchers, offering high-quality models that can be run efficiently on various devices.

Pros

  • Easy to use and integrate into existing projects
  • Supports multiple languages and accents
  • Offers lightweight models suitable for edge devices and mobile applications
  • Provides regular updates and improvements to model performance

Cons

  • Limited customization options for specific use cases
  • May not perform as well as some commercial solutions for certain languages
  • Requires some technical knowledge to set up and use effectively
  • Documentation could be more comprehensive for advanced usage scenarios

Code Examples

  1. Speech-to-Text (STT) example:
import torch
import zipfile
import torchaudio
from omegaconf import OmegaConf

# Load the model
model, decoder, utils = torch.hub.load(repo_or_dir='snakers4/silero-models',
                                       model='silero_stt',
                                       language='en')

# Transcribe audio
wav_path = 'path/to/audio/file.wav'
waveform, sample_rate = torchaudio.load(wav_path)
transcription = model(waveform)
print(transcription)
  1. Text-to-Speech (TTS) example:
import torch

# Load the model
model, _ = torch.hub.load(repo_or_dir='snakers4/silero-models',
                          model='silero_tts',
                          language='en',
                          speaker='en_0')

# Generate speech
text = "Hello, this is a test of text-to-speech synthesis."
audio = model.apply_tts(text=text,
                        speaker='en_0',
                        sample_rate=48000)

# Save the audio
torchaudio.save('output.wav', audio.unsqueeze(0), sample_rate=48000)
  1. Voice Activity Detection (VAD) example:
import torch
import torchaudio

# Load the model
model, utils = torch.hub.load(repo_or_dir='snakers4/silero-models',
                              model='silero_vad',
                              force_reload=True)

# Perform VAD on audio
wav_path = 'path/to/audio/file.wav'
waveform, sample_rate = torchaudio.load(wav_path)
vad_segments = utils.get_speech_timestamps(waveform, model, threshold=0.5)
print(vad_segments)

Getting Started

To get started with Silero Models:

  1. Install the required dependencies:

    pip install torch torchaudio omegaconf
    
  2. Load the desired model using torch.hub.load() as shown in the code examples above.

  3. Process your audio or text data using the loaded model's functions.

  4. For more detailed usage instructions and advanced features, refer to the project's GitHub repository and documentation.

Competitor Comparisons

14,058

Offline speech recognition API for Android, iOS, Raspberry Pi and servers with Python, Java, C# and Node

Pros of vosk-api

  • Supports a wider range of languages and accents
  • Offers offline speech recognition capabilities
  • Provides a more flexible API for integration into various applications

Cons of vosk-api

  • Generally slower processing speed compared to Silero models
  • Requires more computational resources for operation
  • Less focus on lightweight models for edge devices

Code Comparison

vosk-api:

from vosk import Model, KaldiRecognizer
import pyaudio

model = Model("model")
rec = KaldiRecognizer(model, 16000)

p = pyaudio.PyAudio()
stream = p.open(format=pyaudio.paInt16, channels=1, rate=16000, input=True, frames_per_buffer=8000)
stream.start_stream()

while True:
    data = stream.read(4000)
    if rec.AcceptWaveform(data):
        print(rec.Result())

silero-models:

import torch
from silero import models

model, decoder, utils = torch.hub.load(repo_or_dir='snakers4/silero-models', model='silero_stt')

device = torch.device('cpu')
model = model.to(device)

wav_path = 'audio.wav'
transcription = model.transcribe(wav_path)
print(transcription)

Both repositories offer speech recognition capabilities, but they differ in their approach and use cases. vosk-api provides a more comprehensive solution for various languages and offline usage, while silero-models focuses on lightweight, efficient models suitable for edge devices and quick processing.

43,367

🐸💬 - a deep learning toolkit for Text-to-Speech, battle-tested in research and production

Pros of TTS

  • More comprehensive and feature-rich, offering a wider range of TTS models and voice conversion capabilities
  • Active development with frequent updates and a larger community
  • Supports multiple languages and provides pre-trained models for various use cases

Cons of TTS

  • Higher complexity and steeper learning curve compared to Silero Models
  • Requires more computational resources and may have longer inference times
  • Installation process can be more involved, especially for certain features

Code Comparison

TTS:

from TTS.api import TTS

tts = TTS(model_name="tts_models/en/ljspeech/tacotron2-DDC")
tts.tts_to_file(text="Hello world!", file_path="output.wav")

Silero Models:

import torch

model, _ = torch.hub.load(repo_or_dir='snakers4/silero-models', model='silero_tts')
audio = model.apply_tts("Hello world!", speaker='en_0', sample_rate=48000)

Both repositories offer powerful text-to-speech capabilities, but TTS provides a more extensive set of features and models at the cost of increased complexity. Silero Models, on the other hand, offers a simpler and more lightweight solution that may be easier to integrate for basic TTS needs.

DeepSpeech is an open source embedded (offline, on-device) speech-to-text engine which can run in real time on devices ranging from a Raspberry Pi 4 to high power GPU servers.

Pros of DeepSpeech

  • More established project with a larger community and extensive documentation
  • Supports multiple languages and accents out of the box
  • Offers pre-trained models for immediate use

Cons of DeepSpeech

  • Requires more computational resources and has a larger model size
  • Can be more complex to set up and integrate into projects
  • May have slower inference times compared to Silero Models

Code Comparison

DeepSpeech:

import deepspeech
model = deepspeech.Model('path/to/model.pbmm')
text = model.stt(audio)

Silero Models:

import torch
model, decoder, utils = torch.hub.load(repo_or_dir='snakers4/silero-models', model='silero_stt')
text = model(audio)

Both repositories provide speech-to-text functionality, but they differ in implementation and usage. DeepSpeech offers a more comprehensive solution with support for multiple languages, while Silero Models focuses on lightweight, efficient models. DeepSpeech may be better suited for large-scale projects with diverse language requirements, whereas Silero Models could be preferable for applications where speed and resource efficiency are crucial. The choice between the two depends on specific project needs, available computational resources, and desired features.

9,572

End-to-End Speech Processing Toolkit

Pros of ESPnet

  • Comprehensive toolkit with support for various speech processing tasks (ASR, TTS, speech enhancement, etc.)
  • Extensive documentation and active community support
  • Flexible architecture allowing for easy customization and experimentation

Cons of ESPnet

  • Steeper learning curve due to its extensive feature set
  • Potentially higher computational requirements for training and inference
  • More complex setup process compared to simpler alternatives

Code Comparison

ESPnet example (ASR training):

from espnet2.bin.asr_train import main

args = {
    "output_dir": "exp/asr_train",
    "max_epoch": 100,
    "batch_size": 32,
    "accum_grad": 2,
    "use_amp": True
}

main(args)

Silero Models example (ASR inference):

import torch
from silero import models

model, decoder, utils = models.silero_stt(language='en', device='cpu')
audio_path = 'audio.wav'
transcription = model(audio_path)

The ESPnet code showcases its flexibility in training configuration, while Silero Models demonstrates simplicity in inference. ESPnet offers more control over the training process, whereas Silero Models provides a straightforward API for quick deployment.

15,211

kaldi-asr/kaldi is the official location of the Kaldi project.

Pros of Kaldi

  • Comprehensive toolkit with extensive documentation and examples
  • Highly flexible and customizable for various ASR tasks
  • Large community support and active development

Cons of Kaldi

  • Steeper learning curve and more complex setup
  • Requires more computational resources for training and inference
  • Less suitable for quick prototyping or small-scale projects

Code Comparison

Silero-models (Python):

import torch
model, decoder, utils = torch.hub.load(repo_or_dir='snakers4/silero-models', model='silero_stt')
wav_path = 'audio.wav'
transcription = model.transcribe(wav_path)

Kaldi (Shell script):

#!/bin/bash
. ./path.sh
compute-mfcc-feats --config=conf/mfcc.conf scp:data/test/wav.scp ark:- | \
  apply-cmvn-sliding --norm-vars=false --center=true --cmn-window=300 ark:- ark:- | \
  gmm-latgen-faster --max-active=7000 --beam=13.0 --lattice-beam=6.0 --acoustic-scale=0.083333 \
  --allow-partial=true --word-symbol-table=exp/tri4b/graph/words.txt \
  exp/tri4b/final.mdl exp/tri4b/graph/HCLG.fst ark:- ark:- | \
  lattice-best-path --word-symbol-table=exp/tri4b/graph/words.txt ark:- ark,t:transcription.txt
31,926

Facebook AI Research Sequence-to-Sequence Toolkit written in Python.

Pros of fairseq

  • Comprehensive toolkit for sequence modeling tasks
  • Supports a wide range of architectures and tasks
  • Extensive documentation and community support

Cons of fairseq

  • Steeper learning curve due to complexity
  • Requires more computational resources
  • Less focused on specific speech-to-text tasks

Code Comparison

silero-models:

import torch
model, decoder, utils = torch.hub.load(repo_or_dir='snakers4/silero-models', model='silero_stt')

fairseq:

from fairseq.models.wav2vec import Wav2VecModel
model = Wav2VecModel.from_pretrained('/path/to/model')

Key Differences

  • silero-models focuses on speech recognition, while fairseq covers a broader range of sequence modeling tasks
  • silero-models offers simpler integration and usage, while fairseq provides more flexibility and customization options
  • fairseq has a larger codebase and more dependencies, whereas silero-models is more lightweight and easier to deploy

Use Cases

  • Choose silero-models for quick implementation of speech recognition tasks
  • Opt for fairseq when working on complex sequence modeling projects or research requiring extensive customization

Convert Figma logo designs to code with AI

Visual Copilot

Introducing Visual Copilot: A new AI model to turn Figma designs to high quality code using your components.

Try Visual Copilot

README

Mailing list : test Mailing list : test License: CC BY-NC 4.0

PyPI version

header

Silero Models

Our TTS models satisfy the following criteria:

  • Fully end-to-end;
  • Large library of voices;
  • Natural-sounding speech;
  • One-line usage, minimal, portable;
  • Impressively fast on CPU and GPU;
  • For the Russian language - automated stress and homographs;

Installation and Basics

You can basically use our models in 3 flavours:

  • Via PyTorch Hub: torch.hub.load();
  • Via pip: pip install silero and then from silero import silero_tts;
  • Via caching the required models and utils manually and modifying if necessary;

Models are downloaded on demand both by pip and PyTorch Hub. If you need caching, do it manually or via invoking a necessary model once (it will be downloaded to a cache folder). Please see these docs for more information.

PyTorch Hub and pip package are based on the same code. All of the torch.hub.load examples can be used with the pip package via this basic change:

from silero import silero_tts
model, example_text = silero_tts(language='ru',
                                 speaker='v5_ru')
audio = model.apply_tts(text=example_text)

Text-To-Speech

Models and Speakers

All of the provided models are listed in the models.yml file. Any metadata and newer versions will be added there.

V5

V5 models support SSML. Also see Colab examples for main SSML tag usage.

Russian-only models support automated stress and homographs.

IDSpeakersAuto-stress / HomographsLanguageSRColab
v5_ruaidar, baya, kseniya, xenia, eugeneyes / yesru (Russian)8000, 24000, 48000Open In Colab

V5 CIS Base Models

  • All of the below models support 8000, 24000, 48000 sampling rates and contain no auto-stress or homographs;
  • v5_cis_base models assume that proper stress should be added for each word for all languages, i.e. к+ошка;
  • v5_cis_base_nostress models assume that proper stress should be added for each word ONLY for slavic languages (i.e. ru, bel, ukr);
  • All of the below models are published under MIT licence;
  • V5 UTMOS and throughput metrics;
  • V5 models support SSML. Also see Colab examples for main SSML tag usage.
IDSpeakersLanguageColab
v5_cis_base, v5_cis_base_nostressaze_gamataze (Azerbaijani)Open In Colab
v5_cis_base, v5_cis_base_nostresshye_zarahye (Armenian)Open In Colab
v5_cis_base, v5_cis_base_nostressbak_aigul, bak_alfia, bak_alfia2bak (Bashkir)Open In Colab
v5_cis_base, v5_cis_base_nostressbak_miyau, bak_ramiliabak (Bashkir)Open In Colab
v5_cis_base, v5_cis_base_nostressbel_anatoliy, bel_dmitriy, bel_larisabel (Belarus)Open In Colab
v5_cis_base, v5_cis_base_nostresskat_vikakat (Georgian)Open In Colab
v5_cis_base, v5_cis_base_nostresskbd_eduardkbd (Kab.-Cherkes)Open In Colab
v5_cis_base, v5_cis_base_nostresskaz_zhadyra, kaz_zhazirakaz (Kazakh)Open In Colab
v5_cis_base, v5_cis_base_nostressxal_kejilgan, xal_kermenxal (Kalmyk)Open In Colab
v5_cis_base, v5_cis_base_nostresskir_nurgulkir (Kyrgyz)Open In Colab
v5_cis_base, v5_cis_base_nostressmdf_oksanamdf (Moksha)Open In Colab
v5_cis_base, v5_cis_base_nostressall of these speakers, but with ru_ prefixru (Russian)Open In Colab
v5_cis_base, v5_cis_base_nostresstgk_onaoy, tgk_safarhujatgk (Tajik)Open In Colab
v5_cis_base, v5_cis_base_nostresstat_albina, tat_marattat (Tatar)Open In Colab
v5_cis_base, v5_cis_base_nostressudm_bogdanudm (Udmurt)Open In Colab
v5_cis_base, v5_cis_base_nostressuzb_saidauzb (Uzbek)Open In Colab
v5_cis_base, v5_cis_base_nostressukr_igor, ukr_romanukr (Ukrainian)Open In Colab
v5_cis_base, v5_cis_base_nostresskjh_karina, kjh_sibdaykjh (Khakas)Open In Colab
v5_cis_base, v5_cis_base_nostresschv_ekaterinachv (Chuvash)Open In Colab
v5_cis_base, v5_cis_base_nostresserz_alexandrerz (Erzya)Open In Colab
v5_cis_base, v5_cis_base_nostresssah_zinaidasah (Yakut)Open In Colab

V5 CIS Ext Models

  • All of the below models support 8000, 24000, 48000 sampling rates and contain no auto-stress or homographs;
  • v5_cis_ext models assume that proper stress should be added for each word for all languages, i.e. к+ошка;
  • v5_cis_ext_nostress are coming soon;
  • All of the below models are published under CC-NC-BY licence;
  • V5 models support SSML. Also see Colab examples for main SSML tag usage.
IDSpeakersLanguageColab
v5_cis_extkaz_abai, kaz_aidana, kaz_aisha, kaz_bakir, kaz_danarakaz (Kazakh)Open In Colab
v5_cis_extxal_delghir, xal_erdnixal (Kalmyk)Open In Colab
v5_cis_exttat_adiba, tat_alsou, tat_amir, tat_azat, tat_batirtat (Tatar)Open In Colab
v5_cis_exttat_bulat, tat_damir, tat_guzel, tat_ildar, tat_ilgiztat (Tatar)Open In Colab
v5_cis_exttat_karim, tat_mansur, tat_murat, tat_rasima, tat_rustemtat (Tatar)Open In Colab
v5_cis_exttat_timur, tat_zifa, tat_zufar, tat_zulfiyatat (Tatar)Open In Colab
v5_cis_extuzb_anora, uzb_dilnavozuzb (Uzbek)Open In Colab
v5_cis_extukr_kateryna, ukr_lada, ukr_mykyta, ukr_oleksa, ukr_tetianaukr (Ukrainian)Open In Colab
v5_cis_extchv_aihwa, chv_alimachv (Chuvash)Open In Colab

V4

V4 models support SSML. Also see Colab examples for main SSML tag usage.

V4 models: v4_ru, v4_cyrillic, v4_ua, v4_uz, v4_indic
IDSpeakersAuto-stressLanguageSRColab
v4_ruaidar, baya, kseniya, xenia, eugene, randomyesru (Russian)8000, 24000, 48000Open In Colab
v4_cyrillicb_ava, marat_tt, kalmyk_erdni...nocyrillic (Avar, Tatar, Kalmyk, ...)8000, 24000, 48000Open In Colab
v4_uamykyta, randomnoua (Ukrainian)8000, 24000, 48000Open In Colab
v4_uzdilnavoznouz (Uzbek)8000, 24000, 48000Open In Colab
v4_indichindi_male, hindi_female, ..., randomnoindic (Hindi, Telugu, ...)8000, 24000, 48000Open In Colab

V3

V3 models support SSML. Also see Colab examples for main SSML tag usage.

V3 models: v3_en, v3_en_indic, v3_de, v3_es, v3_fr, v3_indic
IDSpeakersAuto-stressLanguageSRColab
v3_enen_0, en_1, ..., en_117, randomnoen (English)8000, 24000, 48000Open In Colab
v3_en_indictamil_female, ..., assamese_male, randomnoen (English)8000, 24000, 48000Open In Colab
v3_deeva_k, ..., karlsson, randomnode (German)8000, 24000, 48000Open In Colab
v3_eses_0, es_1, es_2, randomnoes (Spanish)8000, 24000, 48000Open In Colab
v3_frfr_0, ..., fr_5, randomnofr (French)8000, 24000, 48000Open In Colab
v3_indichindi_male, hindi_female, ..., randomnoindic (Hindi, Telugu, ...)8000, 24000, 48000Open In Colab

Dependencies

Basic dependencies for Colab examples:

  • torch, 1.10+ for v3 models/ 2.0+ for v4 and v5 models;
  • torchaudio, latest version bound to PyTorch should work (required only because models are hosted together with STT, not required for work);
  • omegaconf, latest (can be removed as well, if you do not load all of the configs);

PyTorch

Open In Colab

Open on Torch Hub

# V5
import torch

language = 'ru'
model_id = 'v5_ru'
sample_rate = 48000
speaker = 'xenia'
device = torch.device('cpu')

model, example_text = torch.hub.load(repo_or_dir='snakers4/silero-models',
                                     model='silero_tts',
                                     language=language,
                                     speaker=model_id)
model.to(device)  # gpu or cpu

audio = model.apply_tts(text=example_text,
                        speaker=speaker,
                        sample_rate=sample_rate)

Standalone Use

  • Standalone usage only requires PyTorch 1.12+ and the Python Standard Library;
  • Please see the detailed examples in Colab;
# V5
import os
import torch

device = torch.device('cpu')
torch.set_num_threads(4)
local_file = 'model.pt'

if not os.path.isfile(local_file):
    torch.hub.download_url_to_file('https://models.silero.ai/models/tts/ru/v5_ru.pt',
                                   local_file)  

model = torch.package.PackageImporter(local_file).load_pickle("tts_models", "model")
model.to(device)

example_text = 'Меня зовут Лева Королев. Я из готов. И я уже готов открыть все ваши замки любой сложности!'
sample_rate = 48000
speaker='baya'

audio_paths = model.save_wav(text=example_text,
                             speaker=speaker,
                             sample_rate=sample_rate)

SSML

Check out our TTS Wiki page.

Cyrillic languages v4

To be superseded with v5 model(s) soon.

Supported tokenset: !,-.:?iµöабвгдежзийклмнопрстуфхцчшщъыьэюяёђѓєіјњћќўѳғҕҗҙқҡңҥҫүұҳҷһӏӑӓӕӗәӝӟӥӧөӱӳӵӹ

Speaker_IDLanguageGender
b_avaAvarF
b_bashkirBashkirM
b_bulbBulgarianM
b_bulcBulgarianM
b_cheChechenM
b_cvChuvashM
cv_ekaterinaChuvashF
b_myvErzyaM
b_kalmykKalmykM
b_krcKarachay-BalkarM
kz_M1KazakhM
kz_M2KazakhM
kz_F3KazakhF
kz_F1KazakhF
kz_F2KazakhF
b_kjhKhakasF
b_kpvKomi-ZiryanM
b_lezLezghianM
b_mhrMariF
b_mrjMari HighM
b_nogNogaiF
b_ossOsseticM
b_ruRussianM
b_tatTatarM
marat_ttTatarM
b_tyvTuvinianM
b_udmUdmurtM
b_uzbUzbekM
b_sahYakutM
kalmyk_erdniKalmykM
kalmyk_delghirKalmykF

Indic languages v4

Example

(!!!) All input sentences should be romanized to ISO format using aksharamukha. An example for hindi:

# V3
import torch
from aksharamukha import transliterate

# Loading model
model, example_text = torch.hub.load(repo_or_dir='snakers4/silero-models',
                                     model='silero_tts',
                                     language='indic',
                                     speaker='v4_indic')

orig_text = "प्रसिद्द कबीर अध्येता, पुरुषोत्तम अग्रवाल का यह शोध आलेख, उस रामानंद की खोज करता है"
roman_text = transliterate.process('Devanagari', 'ISO', orig_text)
print(roman_text)

audio = model.apply_tts(roman_text,
                        speaker='hindi_male')

Supported languages

LanguageSpeakersRomanization function
hindihindi_female, hindi_maletransliterate.process('Devanagari', 'ISO', orig_text)
malayalammalayalam_female, malayalam_maletransliterate.process('Malayalam', 'ISO', orig_text)
manipurimanipuri_femaletransliterate.process('Bengali', 'ISO', orig_text)
bengalibengali_female, bengali_maletransliterate.process('Bengali', 'ISO', orig_text)
rajasthanirajasthani_female, rajasthani_femaletransliterate.process('Devanagari', 'ISO', orig_text)
tamiltamil_female, tamil_maletransliterate.process('Tamil', 'ISO', orig_text, pre_options=['TamilTranscribe'])
telugutelugu_female, telugu_maletransliterate.process('Telugu', 'ISO', orig_text)
gujaratigujarati_female, gujarati_maletransliterate.process('Gujarati', 'ISO', orig_text)
kannadakannada_female, kannada_maletransliterate.process('Kannada', 'ISO', orig_text)

Contact

Try our models, create an issue, join our chat, email us, and read the latest news.

Licence

All of the models are published under the main repo license (i.e. CC-NC-BY) except for the base cis-tts models, which are under MIT.

Citations

@misc{Silero Models,
  author = {Silero Team},
  title = {Silero Models: pre-trained text-to-speech models made embarrassingly simple},
  year = {2025},
  publisher = {GitHub},
  journal = {GitHub repository},
  howpublished = {\url{https://github.com/snakers4/silero-models}},
  commit = {insert_some_commit_here},
  email = {hello@silero.ai}
}

Further reading

English

  • STT:

    • Towards an Imagenet Moment For Speech-To-Text - link
    • A Speech-To-Text Practitioners Criticisms of Industry and Academia - link
    • Modern Google-level STT Models Released - link
  • TTS:

    • Multilingual Text-to-Speech Models for Indic Languages - link
    • Our new public speech synthesis in super-high quality, 10x faster and more stable - link
    • High-Quality Text-to-Speech Made Accessible, Simple and Fast - link
  • VAD:

    • One Voice Detector to Rule Them All - link
    • Modern Portable Voice Activity Detector Released - link
  • Text Enhancement:

    • We have published a model for text repunctuation and recapitalization for four languages - link

Chinese

  • STT:
    • 迈向语音识别领域的 ImageNet 时刻 - link
    • 语音领域学术界和工业界的七宗罪 - link

Russian

  • STT

    • OpenAI решили распознавание речи! Разбираемся так ли это … - link
    • Наши сервисы для бесплатного распознавания речи стали лучше и удобнее - link
    • Telegram-бот Silero бесплатно переводит речь в текст - link
    • Бесплатное распознавание речи для всех желающих - link
    • Последние обновления моделей распознавания речи из Silero Models - link
    • Сжимаем трансформеры: простые, универсальные и прикладные способы cделать их компактными и быстрыми - link
    • Ультимативное сравнение систем распознавания речи: Ashmanov, Google, Sber, Silero, Tinkoff, Yandex - link
    • Мы опубликовали современные STT модели сравнимые по качеству с Google - link
    • Понижаем барьеры на вход в распознавание речи - link
    • Огромный открытый датасет русской речи версия 1.0 - link
    • Насколько Быстрой Можно Сделать Систему STT? - link
    • Наша система Speech-To-Text - link
    • Speech-To-Text - link
  • TTS:

    • Делаем быстрый, качественный и доступный синтез на языках России — нужно ваше участие - link
    • Мы решили задачу омографов и ударений в русском языке - link
    • Теперь наш синтез также доступен в виде бота в Телеграме - link
    • Может ли синтез речи обмануть систему биометрической идентификации? - link
    • Теперь наш синтез на 20 языках - link
    • Теперь наш публичный синтез в супер-высоком качестве, в 10 раз быстрее и без детских болячек - link
    • Синтезируем голос бабушки, дедушки и Ленина + новости нашего публичного синтеза - link
    • Мы сделали наш публичный синтез речи еще лучше - link
    • Мы Опубликовали Качественный, Простой, Доступный и Быстрый Синтез Речи - link
  • VAD:

    • Новый релиз публичного детектора голоса Silero VAD v6 - link
    • Наш публичный детектор голоса стал лучше - link
    • А ты используешь VAD? Что это такое и зачем он нужен - link
    • Модели для Детекции Речи, Чисел и Распознавания Языков - link
    • Мы опубликовали современный Voice Activity Detector и не только -link
  • Text Enhancement:

    • Восстановление знаков пунктуации и заглавных букв — теперь и на длинных текстах - link
    • Мы опубликовали модель, расставляющую знаки препинания и заглавные буквы в тексте на четырех языках - link