Orca Text-to-Speech engine for web browsers (via WebAssembly)
npm install @picovoice/orca-webMade in Vancouver, Canada by Picovoice
Orca is an on-device streaming text-to-speech engine that is designed for use with LLMs, enabling zero-latency
voice assistants. Orca is:
- Private; All speech synthesis runs locally.
- Cross-Platform:
- Linux (x86_64), macOS (x86_64, arm64), and Windows (x86_64)
- Android and iOS
- Chrome, Safari, Firefox, and Edge
- Raspberry Pi (3, 4, 5)
- Chrome / Edge
- Firefox
- Safari
Orca Web Binding uses SharedArrayBuffer
for processing speaker diarization.
Include the following headers in the response to enable the use of SharedArrayBuffers:
```
Cross-Origin-Opener-Policy: same-origin
Cross-Origin-Embedder-Policy: require-corp
Refer to our Web demo for an example on creating a server with the corresponding response headers.
Browsers that lack support for SharedArrayBuffers or required headers will fall back to using standard ArrayBuffers,
which disables multithreaded performance.
IndexedDB is required to use Orca in a worker thread. Browsers without IndexedDB supportOrca
(i.e. Firefox Incognito Mode) should use in the main thread.
Multi-threading is only enabled for Orca when using on a web worker.
Using Yarn:
`console`
yarn add @picovoice/orca-web
or using npm:
`console`
npm install --save @picovoice/orca-web
Orca requires a valid Picovoice AccessKey at initialization. AccessKey acts as your credentials when using OrcaAccessKey
SDKs.
You can get your for free. Make sure to keep your AccessKey secret.AccessKey
Signup or Login to Picovoice Console to get your .
For the web packages, there are two methods to initialize Orca.
NOTE: Due to modern browser limitations of using a file URL, this method does __not__ work if used without hosting a
server.
This method fetches the model file from the public directory
and feeds it to Orca. Copy a model file for the desired language and voice into the public
directory:
`console`
cp ${ORCA_MODEL_FILE} ${PATH_TO_PUBLIC_DIRECTORY}
NOTE: This method works without hosting a server, but increases the size of the model file roughly by 33%.
This method uses a base64 string of the model file and feeds it to Orca. Use the built-in script pvbase64 to
base64 your model file:
`console`
npx pvbase64 -i ${ORCA_MODEL_FILE} -o ${OUTPUT_DIRECTORY}/${MODEL_NAME}.js
The output will be a js file which you can import into any file of your project. For detailed information
about pvbase64,
run:
`console`
npx pvbase64 -h
Orca saves and caches your model file in IndexedDB to be used by WebAssembly. Use a different customWritePath variableforceWrite
to hold multiple models and set the value to true to force re-save a model file.
Either base64 or publicPath must be set to instantiate Orca. If both are set, Orca will use the base64 model.
`typescript
const orcaModel = {
publicPath: "${MODEL_RELATIVE_PATH}",
// or
base64: "${MODEL_BASE64_STRING}",
// Optionals
customWritePath: "orca_model",
forceWrite: false,
version: 1,
}
`
Create an instance of Orca on the main thread:
`typescript`
const orca = await Orca.create(
"${ACCESS_KEY}",
orcaModel
);
Or create an instance of Orca in a worker thread:
`typescript`
const orca = await OrcaWorker.create(
"${ACCESS_KEY}",
orcaModel
);
Orca supports two modes of operation: streaming and single synthesis.
In the streaming synthesis mode, Orca processes an incoming text stream in real-time and generates audio in parallel.
In the single synthesis mode, the complete text needs to be known in advance and is synthesized in a single call to
the Orca engine.
Orca allows the embedding of custom pronunciations in the text via the syntax: {word|pronunciation}. The pronunciation
is expressed in ARPAbet phonemes, for example:
- "This is a {custom|K AH S T AH M} pronunciation"
- "{read|R IY D} this as {read|R EH D}, please."
- "I {live|L IH V} in {Sevilla|S EH V IY Y AH}. We have great {live|L AY V} sports!"
To obtain the complete set of valid characters, call .validCharacters. To retrieve the maximum number of.maxCharacterLimit
characters allowed, call . The sample rate of the generated Int16Array is .sampleRate.
#### Streaming Synthesis
To use streaming synthesis, call streamOpen to create an OrcaStream object.
`typescript`
const orcaStream = await orca.streamOpen();
Then, call synthesize on orcaStream to generate speech from a stream of text:
`typescript
function* textStream(): IterableIterator
... // yield text chunks e.g. from an LLM response
}
for (const textChunk of textStream()) {
const pcm = await orcaStream.synthesize(textChunk);
if (pcm !== null) {
// handle pcm
}
}
`
The OrcaStream object buffers input text until there is enough to generate audio. If there is not enough text tonull
generate
audio, is returned.
When done, call flush to synthesize any remaining text, and close to delete the orcaStream object.
`typescript
const flushedPcm = orcaStream.flush();
if (flushedPcm !== null) {
// handle pcm
}
orcaStream.close();
`
#### Single Synthesis
To use single synthesis, simply call synthesize directly on the Orca instance. The synthesize function will sendInt16Array
the text to the engine and return the speech audio data as an as well as
the alignments metadata.
`typescript`
const { pcm, alignments } = await orca.synthesize("${TEXT}");
Orca allows for additional arguments to control the synthesized speech.
These can be provided to streamOpen or one of the single mode synthesize methods:
- speechRate: Controls the speed of the generated speech. Valid values are within [0.7, 1.3]. A higher value produces1.0
speech that is faster, and a lower value produces speech that is slower. The default value is .
`typescript
const synthesizeParams = {
speechRate: 1.3,
};
// Streaming synthesis
const OrcaStream = await orca.streamOpen(synthesizeParams);
// Single synthesis
const result = await orca.synthesize("${TEXT}", synthesizeParams);
`
Along with the raw PCM or saved audio file, Orca returns metadata for the synthesized audio in single synthesis mode.
The OrcaAlignment object has the following properties:
- Word: String representation of the word.
- Start Time: Indicates when the word started in the synthesized audio. Value is in seconds.
- End Time: Indicates when the word ended in the synthesized audio. Value is in seconds.
- Phonemes: An array of OrcaPhoneme objects.
The OrcaPhoneme object has the following properties:
- Phoneme: String representation of the phoneme.
- Start Time: Indicates when the phoneme started in the synthesized audio. Value is in seconds.
- End Time: Indicates when the phoneme ended in the synthesized audio. Value is in seconds.
Clean up used resources by Orca or OrcaWorker:
`typescript`
await orca.release();
Terminate OrcaWorker instance:
`typescript``
await orca.terminate();
For example usage refer to our Web demo application.