AI voices

> Welcome > AI voices

CЕРВЕР СИНТЕЗА РЕЧИ

Table of Contents

Speech synthesis is needed to convert text into spoken language.

  • Blind and visually impaired people. It allows them to listen to texts, books, websites, and documents.
  • For people with speech impairments. Speech synthesis can “speak” for them (via special devices).
  • Voice assistants (Siri, Alexa, and others).
  • Call centers.
  • To assist with foreign language learning.
  • For dubbing and voice-over of any content.
  • For creating audiobooks.

How is speech synthesis created in a new language?

MultiVoice.AI enlists the help of a philologist and a speaker to create a new language.
The philologist’s job is to create a dictionary of specialized terms and abbreviations for the language. The philologist also reviews the work of the speaker. The speaker must speak audio files. The speaker must have clear diction and pronounce all cases and endings clearly and distinctly. The philologist must check each phrase spoken by the speaker for compliance with the linguistic rules of the language (the speaker and philologist can work remotely from home). MultiVoice.AI’s specialized computer program then processes the above information, and after some time, a computer file is created—an electronic model of the language. Testing then takes place—the philologist verifies the accuracy of the resulting electronic model. If the preliminary work by the philologist and speaker is high-quality and accurate, the electronic model will be able to synthesize and pronounce any text. The speaker must record approximately 100 hours of HIGH-QUALITY audio for the electronic model to function properly. Other voices can then be added to the electronic language model, allowing the speech synthesizer to pronounce phrases. Creating speech synthesis for a new language takes an average of 3-5 months (depending on various features). Creating speech synthesis for a new language involves four people: a philologist, a voice-over artist, a technical specialist from MultiVoice.AI, and a project manager from MultiVoice.AI.

What do we offer?

1. We are ready to create speech synthesizers for any minority languages ​​in the world. Minority languages ​​are those spoken by fewer than 10 million people. Once a language model is created, it will be possible to voice/dub any video and audio content with up to 24 different voices.
2. Creating a new language takes approximately 3-5 months and costs approximately 10-12 thousand euros (including a linguist and a voice actor for 3-5 months of work, including taxes). MultiVoice.AI will provide its resources free of charge.
3. Using the MultiVoice.AI API, you can integrate speech synthesis into your application.

What we have already done:

MultiVoice.AI has created its own speech synthesis server for Latvian, Russian, and Ukrainian. Ten male, ten female, and four child voices are available for each language, for a total of 24 different voices for each language. Speech synthesis is available via an API (for developers) or through the dedicated Dubbing Studio app (available for Windows and Mac).

You can test speech synthesis in Latvian, Russian, and Ukrainian using this link (using 24 voices for each language).

What languages ​​are in development?

MultiVoice.AI is currently working on technical development and developing speech synthesis for the following languages: Lithuanian, Estonian, English, Chinese, Japanese, German, French, Spanish, Korean, Arabic, Danish, Italian, Polish, and Portuguese. Each language will feature 24 different voices.

3 TV channels

Our speech synthesis is already used by three TV channels in Latvia for dubbing or voice-over translation of films.

10 male voices

MultiVoice.Ai has created its own speech synthesis server for Russian, Latvian, and Ukrainian. Twelve more languages ​​are in development.

10 female voices

There are 24 unique voices available for each language: 10 male, 10 female, and 4 child voices.

4 children's voices

Children's voices can be used to voice or dub animated films.

Advantages of Dubbing Studio:

Questions and Answers

Yes, you can. See the API section of our website.

We independently develop speech synthesis models for various languages. Our first model is a speech synthesis model for Latvian, Russian, and Ukrainian, using 24 different voices (10 male, 10 female, and 4 child voices). Our team independently creates and updates the dataset for the speech synthesis model monthly. We have hired Latvian, Russian, and Ukrainian philologists to validate and refine our language model. We believe that our speech synthesis model for Latvian, Russian, and Ukrainian is as good as AI speech synthesis from Google, Amazon, Event Labs, and Tilde (for Latvian).

You can test speech synthesis in Latvian, Russian and Ukrainian (using 24 voices) using this link.

Reviews

Our clients

Movify.CY Ltd.
Reg.nr. ΗΕ461401
Vat 60077645I
Agia Zoni, 30a, Fantaros Court,
Flat/Office 303, 3027, Limassol,
Cyprus.