The original version, here: https://github.com/Migushthe2nd/MsEdgeTTS Was undermaintained and had dependency issues, so I forked the repo and fixed them.
An simple Azure Speech Service module that uses the Microsoft Edge Read Aloud API.
Only supports speak
, voice
, and prosody
element types. The following is the default SSML object:
<speak version="1.0" xmlns="http://www.w3.org/2001/10/synthesis" xmlns:mstts="https://www.w3.org/2001/mstts"
xml:lang="${this._voiceLang}">
<voice name="${voiceName}">
<prosody rate="${rate}" pitch="${pitch}" volume="${volume}">
${input}
</prosody>
</voice>
</speak>
Documentation on the SSML format can be found here . All supported audio formats can be found here.
Make sure to escape/sanitize your user's input! Use a library like xml-escape.
import { MsEdgeTTS, OUTPUT_FORMAT } from "msedge-tts";
const tts = new MsEdgeTTS();
await tts.setMetadata(
"en-IE-ConnorNeural",
OUTPUT_FORMAT.WEBM_24KHZ_16BIT_MONO_OPUS
);
const readable = tts.toStream("Hi, how are you?");
readable.on("data", (data) => {
console.log("DATA RECEIVED", data);
// raw audio file data
});
readable.on("close", () => {
console.log("STREAM CLOSED");
});
import { MsEdgeTTS, OUTPUT_FORMAT } from "msedge-tts";
(async () => {
const tts = new MsEdgeTTS();
await tts.setMetadata(
"en-US-AriaNeural",
OUTPUT_FORMAT.WEBM_24KHZ_16BIT_MONO_OPUS
);
const filePath = await tts.toFile("./example_audio.webm", "Hi, how are you?");
})();
import { MsEdgeTTS, OUTPUT_FORMAT } from "msedge-tts";
(async () => {
const tts = new MsEdgeTTS();
await tts.setMetadata(
"en-US-AriaNeural",
OUTPUT_FORMAT.WEBM_24KHZ_16BIT_MONO_OPUS
);
const filePath = await tts.toFile(
"./example_audio.webm",
"Hi, how are you?",
{ rate: 0.5, pitch: "+200Hz" }
);
})();
Use a custom http.Agent implementation like https-proxy-agent or socks-proxy-agent.
import { SocksProxyAgent } from "socks-proxy-agent";
(async () => {
const agent = new SocksProxyAgent(
"socks://your-name%40gmail.com:abcdef12345124@br41.nordvpn.com"
);
const tts = new MsEdgeTTS(agent);
await tts.setMetadata(
"en-US-AriaNeural",
OUTPUT_FORMAT.WEBM_24KHZ_16BIT_MONO_OPUS
);
const filePath = await tts.toFile("./example_audio.webm", "Hi, how are you?");
})();
For the full documentation check out the API Documentation.
This library only supports promises.