You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
description: 'Learn about the different ways to use our APIs, including cloud services and on-prem containers.'
3
+
description: "Learn about the different ways to use our APIs, including cloud services and on-prem containers."
4
4
---
5
5
6
6
# Overview
@@ -10,6 +10,7 @@ description: 'Learn about the different ways to use our APIs, including cloud se
10
10
Leverage Speechmatics’ cloud services for easy, scalable, and fully managed speech-to-text and translation capabilities.
11
11
12
12
The best way to get started using Speechmatics' cloud services is:
13
+
13
14
- Create an account in our [Portal](https://portal.speechmatics.com/)
14
15
- Check out our [Realtime Transcription](/speech-to-text/realtime/quickstart.mdx)
15
16
- Check out our [Batch Transcription](/speech-to-text/batch/quickstart.mdx)
@@ -22,30 +23,29 @@ Deploy Speechmatics services in your own environment using containers. This opti
22
23
-[Language ID Container](/deployments/container/language-id): Identify the language spoken in your audio using the Language ID container.
23
24
-[Translation Container](/deployments/container/gpu-translation): Translate audio from one language to another using the Translation container.
24
25
25
-
26
26
## Feature Availability
27
27
28
28
Feature availability varies depending on the deployment method you choose. Below is a table summarizing the speech to text feature availability for each deployment method and processing mode.
Copy file name to clipboardExpand all lines: docs/deployments/virtual-appliance/administration/adding-languages.mdx
+13-3Lines changed: 13 additions & 3 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -4,6 +4,10 @@ description: Add languages to a Virtual Appliance deployment
4
4
5
5
# Adding Languages
6
6
7
+
:::warning
8
+
As of 08 October 2025 we have moved our public containers from artifactory to Azure Container Registry (ACR), please [reach out to Support](https://support.speechmatics.com) to update your credentials.
9
+
:::
10
+
7
11
:::note
8
12
Adding languages is currently only supported for `batch` mode, if you need extra realtime languages please [reach out to Support](https://support.speechmatics.com).
9
13
:::
@@ -15,21 +19,27 @@ Log into the appliance using SSH with the username `smadmin`, as described in [R
15
19
You will need to know which version of the images are compatible with your appliance, this will be documented here, but until then [reach out to Support](https://support.speechmatics.com).
16
20
Language codes are the standard two-letter ISO codes used in configuring transcription, for example: "de", "fr".
17
21
22
+
Appliance versions up to and including `6.2.1` will pull from our artifactory repository by default, as of 08 October 2025 we will deprecate public access to this repository (see above). If running an appliance version `6.2.1` or earlier, you will need to configure your appliance to pull from the ACR repository instead.
Access to the Speechmatics repository is needed to download new languages, these credentials can be set in the environment but need to be base64 encoded, alternately if unset the user will be prompted for the username and password after running the configure languages script. If you need access to the Speechmatics container repository please [reach out to Support](https://support.speechmatics.com).
Your API key must be provided in the WebSocket connection request header. For example:
36
+
### Server-side using header + API key
37
+
38
+
For server-side calls, your API key must be provided in the header of the [Websocket connection request](/api-ref/realtime-transcription-websocket#handshake-responses).
39
+
40
+
### Client-side using temporary key
41
+
42
+
For client-side calls, we recommend using [temporary keys](#temporary-keys) (JWTs) to authenticate your requests. These can be passed as a query param to the [Websocket connection request](/api-ref/realtime-transcription-websocket#handshake-responses) URL.
43
+
44
+
For example:
37
45
38
46
```bash
39
-
wss://eu2.rt.speechmatics.com/v2
47
+
wss://eu2.rt.speechmatics.com/v2?jwt=$TEMP_KEY
40
48
```
41
49
42
50
</TabItem>
@@ -54,7 +62,7 @@ Speechmatics Batch SaaS supports the following endpoints for production use:
54
62
| Enterprise | EU | EU2 |**eu2.asr.api.speechmatics.com**|
55
63
| All | US | US1 |**us.asr.api.speechmatics.com**<br/>**us1.asr.api.speechmatics.com**|
56
64
| Enterprise | US | US2 |**us2.asr.api.speechmatics.com**|
57
-
|Enterprise| AU | AU1 |**au1.asr.api.speechmatics.com**|
65
+
|All| AU | AU1 |**au1.asr.api.speechmatics.com**|
58
66
59
67
Speechmatics Real-Time SaaS supports the following endpoints for production use:
60
68
@@ -64,20 +72,22 @@ Speechmatics Real-Time SaaS supports the following endpoints for production use:
64
72
| All | EU | EU1 | neu.rt.speechmatics.com |
65
73
| All | US | US1 | wus.rt.speechmatics.com |
66
74
67
-
Note that for Enterprise customers with access to the Portal, only jobs created in the EU1 environment will appear in the [Jobs list](https://portal.speechmatics.com/jobs).
75
+
All production environments are active and highly available. You can use multiple environments to balance requests or provide a failover in the event of disruption to one environment.
68
76
69
-
All production environments are active and highly available. Multiple environments can be used to balance requests or provide a failover in the event of disruption to one environment.
77
+
Jobs are created in the environment corresponding to the endpoint used. You must use the same endpoint for all requests relating to a specific job.
70
78
71
-
Note that jobs are created in the environment corresponding to the endpoint used. You must use the same endpoint for all requests relating to a specific job.
79
+
If you attempt to use an endpoint for a region you are not contracted to use, that request will be unsuccessful. If you want to use a different region, please contact your account manager.
72
80
73
-
If you attempt to use an endpoint for a region you are not contracted to use, that request will be unsuccessful. If you want to use a different region, please contact your Account Manager.
81
+
:::warning
82
+
The EU2 and US2 environments are provided for enterprise customer high availability and failover purposes only. Jobs created in these environments will not be visible in the portal.
83
+
:::
74
84
75
85
## Temporary keys
76
86
77
87
Speechmatics allows you to generate temporary keys to authenticate your requests instead of your long-lived API key. This can improve end-user experience by reducing latency, and can also reduce development effort and complexity.
78
88
79
89
:::info
80
-
If you are an enterprise customer and would like to use temporary keys, reach out to [Support](https://support.speechmatics.com) or speak to your Account Manager.
90
+
If you are an enterprise customer and would like to use temporary keys, reach out to [support](https://support.speechmatics.com) or speak to your account manager.
81
91
:::
82
92
83
93
### Real-time transcription
@@ -117,6 +127,15 @@ curl -L -X POST "https://mp.speechmatics.com/v1/api_keys?type=batch" \
117
127
-d '{"ttl": 60, "client_ref": "USER123"}'
118
128
```
119
129
130
+
The response JSON looks like this:
131
+
132
+
```json
133
+
{
134
+
"apikey_id": null,
135
+
"key_value": "eyJhbG..."
136
+
}
137
+
```
138
+
120
139
To authorize a request using a temporary key, simply use it in place of the API key in the Authorization header.
121
140
122
141
For example, if you created a temporary key associated with a given `client_ref` you can retrieve those jobs as follows:
Copy file name to clipboardExpand all lines: docs/speech-to-text/batch/batch-diarization.mdx
+1-5Lines changed: 1 addition & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -52,7 +52,7 @@ The feature is disabled by default. To enable speaker diarization, `diarization`
52
52
When diarization is enabled, each `word` and `punctuation` object in the transcript includes a `speaker` property that identifies who spoke it. There are two types of labels:
53
53
54
54
-`S#` – S stands for speaker, and `#` is a sequential number identifying each speaker. S1 appears first in the results, followed by S2, S3, and so on.
55
-
-`UU` – Used when the speaker cannot be identified or diarization is not applied (i.e. running batch mode on CPU operating points), for example, if background noise is transcribed as speech but no speaker can be determined.
55
+
-`UU` – Used when the speaker cannot be identified or diarization is not applied, for example, if background noise is transcribed as speech but no speaker can be determined.
56
56
57
57
```json
58
58
"results": [
@@ -184,10 +184,6 @@ This adjustment only works when punctuation is enabled. Disabling punctuation vi
184
184
185
185
Adjusting punctuation sensitivity can also affect how accurately speakers are identified.
186
186
187
-
### Speaker diarization Timeout
188
-
189
-
Speaker diarization will time out if it takes too long to run for a particular audio file. Currently, the timeout is set to 5 minutes or 0.5 \* the audio duration, whichever is longer. For example, with a 2 hour audio file, the timeout is 1 hour. If a timeout happens, the transcript will still be returned and all speaker labels in the output will be labelled as UU.
190
-
191
187
### Speaker change (legacy)
192
188
193
189
The speaker change detection feature was removed in July 2024. The `speaker_change` and `channel_and_speaker_change` parameters are no longer supported. Use the [speaker diarization](#speaker-diarization) feature for speaker labeling.
0 commit comments