diff --git a/fern/assets/img/faq/api-key.png b/fern/assets/img/faq/api-key.png new file mode 100644 index 00000000..0049d0f6 Binary files /dev/null and b/fern/assets/img/faq/api-key.png differ diff --git a/fern/assets/img/faq/autopay.png b/fern/assets/img/faq/autopay.png new file mode 100644 index 00000000..5dbd901f Binary files /dev/null and b/fern/assets/img/faq/autopay.png differ diff --git a/fern/assets/img/faq/error.png b/fern/assets/img/faq/error.png new file mode 100644 index 00000000..1097ee0f Binary files /dev/null and b/fern/assets/img/faq/error.png differ diff --git a/fern/assets/img/faq/error2.jpg b/fern/assets/img/faq/error2.jpg new file mode 100644 index 00000000..fd015773 Binary files /dev/null and b/fern/assets/img/faq/error2.jpg differ diff --git a/fern/assets/img/faq/playground.png b/fern/assets/img/faq/playground.png new file mode 100644 index 00000000..a2a5ff5a Binary files /dev/null and b/fern/assets/img/faq/playground.png differ diff --git a/fern/assets/img/faq/playground2.png b/fern/assets/img/faq/playground2.png new file mode 100644 index 00000000..44484c66 Binary files /dev/null and b/fern/assets/img/faq/playground2.png differ diff --git a/fern/assets/img/faq/support.png b/fern/assets/img/faq/support.png new file mode 100644 index 00000000..c1c7d5fc Binary files /dev/null and b/fern/assets/img/faq/support.png differ diff --git a/fern/assets/img/faq/support2.png b/fern/assets/img/faq/support2.png new file mode 100644 index 00000000..fdf43b1e Binary files /dev/null and b/fern/assets/img/faq/support2.png differ diff --git a/fern/docs.yml b/fern/docs.yml index b950af75..6f5bd14e 100644 --- a/fern/docs.yml +++ b/fern/docs.yml @@ -28,7 +28,7 @@ tabs: skip-slug: true faq: display-name: FAQ - href: https://support.assemblyai.com/ + slug: faq playground: display-name: Playground href: https://www.assemblyai.com/playground @@ -725,6 +725,449 @@ navigation: path: pages/05-guides/transcribing-an-audio-file.mdx hidden: true - tab: faq + layout: + - section: Overview + path: pages/faq/index.mdx + skip-slug: true + contents: + - section: Account, Billing & Payments + skip-slug: true + contents: + - page: Can an account have more than one API key? + path: pages/faq/accounts-billing-pricing/can-an-account-have-more-than-one-api-key.mdx + slug: /can-an-account-have-more-than-one-api-key + - page: Can I sign up for free? + path: pages/faq/accounts-billing-pricing/can-i-sign-up-for-free.mdx + slug: /can-i-sign-up-for-free + - page: Do my free credits expire? + path: pages/faq/accounts-billing-pricing/do-my-free-credits-expire.mdx + slug: /do-my-free-credits-expire + - page: Do you have any type of special pricing for non-profit companies? + path: pages/faq/accounts-billing-pricing/do-you-have-any-type-of-special-pricing-for-non-profit-companies.mdx + slug: /do-you-have-any-type-of-special-pricing-for-non-profit-companies + - page: Do you offer a student discount? + path: pages/faq/accounts-billing-pricing/do-you-offer-a-student-discount.mdx + slug: /do-you-offer-a-student-discount + - page: Do you offer special pricing for start-ups? + path: pages/faq/accounts-billing-pricing/do-you-offer-special-pricing-for-start-ups.mdx + slug: /do-you-offer-special-pricing-for-start-ups + - page: Do you offer special pricing for YCombinator companies? + path: pages/faq/accounts-billing-pricing/do-you-offer-special-pricing-for-ycombinator-companies.mdx + slug: /do-you-offer-special-pricing-for-ycombinator-companies + - page: Do you offer volume discounts? + path: pages/faq/accounts-billing-pricing/do-you-offer-volume-discounts.mdx + slug: /do-you-offer-volume-discounts + - page: How do I add developers to my account? + path: pages/faq/accounts-billing-pricing/how-do-i-add-developers-to-my-team.mdx + slug: /how-do-i-add-developers-to-my-team + - page: How do I change my account email address? + path: pages/faq/accounts-billing-pricing/how-do-i-change-my-account-email-address.mdx + slug: /how-do-i-change-my-account-email-address + - page: How do I update company information on my invoices? + path: pages/faq/accounts-billing-pricing/how-do-i-update-company-information-on-my-invoices.mdx + slug: /how-do-i-update-company-information-on-my-invoices + - page: How does Auto-pay work? + path: pages/faq/accounts-billing-pricing/how-does-auto-pay-work.mdx + slug: /how-does-auto-pay-work + - page: How does pricing work? + path: pages/faq/accounts-billing-pricing/how-does-pricing-work.mdx + slug: /how-does-pricing-work + - page: How does the concurrency limit work for transcription requests? + path: pages/faq/accounts-billing-pricing/how-does-the-concurrency-limit-work-for-transcription-requests.mdx + slug: /how-does-the-concurrency-limit-work-for-transcription-requests + - page: How often does the Usage and Spend Information in my Dashboard update? + path: pages/faq/accounts-billing-pricing/how-often-does-the-usage-and-spend-information-in-my-dashboard-update-.mdx + slug: /how-often-does-the-usage-and-spend-information-in-my-dashboard-update- + - page: How to Delete Your Account + path: pages/faq/accounts-billing-pricing/how-to-delete-your-account.mdx + slug: /how-to-delete-your-account + - page: How to an get AssemblyAI API Key + path: pages/faq/accounts-billing-pricing/how-to-get-your-api-key.mdx + slug: /how-to-get-your-api-key + - page: How to Share Account Access with Team Members + path: pages/faq/accounts-billing-pricing/how-to-share-account-access-with-team-members.mdx + slug: /how-to-share-account-access-with-team-members + - page: Is AssemblyAI available on the AWS Marketplace? + path: pages/faq/accounts-billing-pricing/is-assemblyai-available-on-the-aws-marketplace.mdx + slug: /is-assemblyai-available-on-the-aws-marketplace + - page: What are my concurrency limits? + path: pages/faq/accounts-billing-pricing/what-are-my-concurrency-limits.mdx + slug: /what-are-my-concurrency-limits + - page: What happens if I reach my concurrency limit? + path: pages/faq/accounts-billing-pricing/what-happens-if-i-reach-my-concurrency-limit.mdx + slug: /what-happens-if-i-reach-my-concurrency-limit + - page: What payment methods do you accept? + path: pages/faq/accounts-billing-pricing/what-payment-methods-do-you-accept.mdx + slug: /what-payment-methods-do-you-accept + - section: Privacy & Security + skip-slug: true + contents: + - page: Does AssemblyAI use encryption? + path: pages/faq/privacy-security/are-audio-and-text-files-encrypted.mdx + slug: /are-audio-and-text-files-encrypted + - page: Are files submitted to the API used for model training? + path: pages/faq/privacy-security/are-files-submitted-to-the-api-used-for-model-training.mdx + slug: /are-files-submitted-to-the-api-used-for-model-training + - page: Are you GDPR compliant? + path: pages/faq/privacy-security/are-you-gdpr-compliant.mdx + slug: /are-you-gdpr-compliant + - page: Are you HIPAA compliant? + path: pages/faq/privacy-security/are-you-hipaa-compliant.mdx + slug: /are-you-hipaa-compliant + - page: Can I sign a DPA agreement with AssemblyAI? + path: pages/faq/privacy-security/can-i-sign-a-dpa-agreement-with-assemblyai.mdx + slug: /can-i-sign-a-dpa-agreement-with-assemblyai + - page: Can you provide a copy of your most recent penetration test executive summary? + path: pages/faq/privacy-security/can-you-provide-a-copy-of-your-most-recent-penetration-test-executive-summary.mdx + slug: /can-you-provide-a-copy-of-your-most-recent-penetration-test-executive-summary + - page: Can you provide a recent vulnerability scan? + path: pages/faq/privacy-security/can-you-provide-a-recent-vulnerability-scan.mdx + slug: /can-you-provide-a-recent-vulnerability-scan + - page: Can I sign a BAA with AssemblyAI? + path: pages/faq/privacy-security/can-you-sign-a-baa.mdx + slug: /can-you-sign-a-baa + - page: Do customers own the system outputs from AssemblyAI? + path: pages/faq/privacy-security/do-customers-own-the-system-outputs-from-assemblyai.mdx + slug: /do-customers-own-the-system-outputs-from-assemblyai + - page: Do you have a formal risk assessment policy or process? + path: pages/faq/privacy-security/do-you-have-a-formal-risk-assessment-policy-or-process.mdx + slug: /do-you-have-a-formal-risk-assessment-policy-or-process + - page: Do you have documented information security policies? If so, how frequently are they updated? + path: pages/faq/privacy-security/do-you-have-documented-information-security-policies-if-so-how-frequently-are-they-updated.mdx + slug: /do-you-have-documented-information-security-policies-if-so-how-frequently-are-they-updated + - page: Do you have SOC2 certification? + path: pages/faq/privacy-security/do-you-have-soc2-certification.mdx + slug: /do-you-have-soc2-certification + - page: Do you have supporting documentation showing your alignment with various privacy frameworks? + path: pages/faq/privacy-security/do-you-have-supporting-documentation-showing-your-alignment-with-various-privacy-frameworks.mdx + slug: /do-you-have-supporting-documentation-showing-your-alignment-with-various-privacy-frameworks + - page: Do you offer EU Data Residency? + path: pages/faq/privacy-security/do-you-offer-eu-data-residency.mdx + slug: /do-you-offer-eu-data-residency + - page: Do you offer servers in the EU? + path: pages/faq/privacy-security/do-you-offer-servers-in-the-eu.mdx + slug: /do-you-offer-servers-in-the-eu + - page: Do you support SAML in your product? + path: pages/faq/privacy-security/do-you-support-saml-in-your-product.mdx + slug: /do-you-support-saml-in-your-product + - page: Does AssemblyAI classify information assets? + path: pages/faq/privacy-security/does-assemblyai-classify-information-assets.mdx + slug: /does-assemblyai-classify-information-assets + - page: Does AssemblyAI offer Zero Data Retention? + path: pages/faq/privacy-security/does-assemblyai-offer-zero-data-retention.mdx + slug: /does-assemblyai-offer-zero-data-retention + - page: Does AssemblyAI utilize an anti-virus/anti-malware solution across all relevant infrastructure (workstations and servers), and are appropriate response capabilities deployed to respond to alerts? + path: pages/faq/privacy-security/does-assemblyai-utilize-an-anti-virus-anti-malware-solution-across-all-relevant-infrastructure-workstations-and-servers-and-are-appropriate-response-capabilities-deployed-to-respond-to-ale.mdx + slug: /does-assemblyai-utilize-an-anti-virus-anti-malware-solution-across-all-relevant-infrastructure-workstations-and-servers-and-are-appropriate-response-capabilities-deployed-to-respond-to-ale + - page: How are incidents escalated within your organization? + path: pages/faq/privacy-security/how-are-incidents-escalated-within-your-organization.mdx + slug: /how-are-incidents-escalated-within-your-organization + - page: How do we securely use your service? + path: pages/faq/privacy-security/how-do-we-securely-use-your-service.mdx + slug: /how-do-we-securely-use-your-service + - page: How do you ensure production code is secure? + path: pages/faq/privacy-security/how-do-you-ensure-production-code-is-secure.mdx + slug: /how-do-you-ensure-production-code-is-secure + - page: How does AssemblyAI back up inputs to its system? + path: pages/faq/privacy-security/how-does-assemblyai-back-up-inputs-to-its-system.mdx + slug: /how-does-assemblyai-back-up-inputs-to-its-system + - page: How does AssemblyAI encrypt data at rest and in transit? + path: pages/faq/privacy-security/how-does-assemblyai-encrypt-data-at-rest-and-in-transit.mdx + slug: /how-does-assemblyai-encrypt-data-at-rest-and-in-transit + - page: How does AssemblyAI segment its environment? + path: pages/faq/privacy-security/how-does-assemblyai-segment-its-environment.mdx + slug: /how-does-assemblyai-segment-its-environment + - page: How long are files submitted to the API retained? + path: pages/faq/privacy-security/how-long-are-files-submitted-to-the-api-retained.mdx + slug: /how-long-are-files-submitted-to-the-api-retained + - page: How long are outputs retained? + path: pages/faq/privacy-security/how-long-are-outputs-retained.mdx + slug: /how-long-are-outputs-retained + - page: How to Access AssemblyAI's Security Reports + path: pages/faq/privacy-security/how-to-access-assemblyai-s-security-reports.mdx + slug: /how-to-access-assemblyai-s-security-reports + - page: How to Opt Out of Data Sharing for Model Training + path: pages/faq/privacy-security/how-to-opt-out-of-data-sharing-for-model-training.mdx + slug: /how-to-opt-out-of-data-sharing-for-model-training + - page: Is multi-factor authentication enforced for all access to scoped systems and data? + path: pages/faq/privacy-security/is-multi-factor-authentication-enforced-for-all-access-to-scoped-systems-and-data.mdx + slug: /is-multi-factor-authentication-enforced-for-all-access-to-scoped-systems-and-data + - page: Does AssemblyAI have a documented process for reviewing and approving third-party service providers? + path: pages/faq/privacy-security/is-there-a-documented-process-for-reviewing-and-approving-third-party-service-providers.mdx + slug: /is-there-a-documented-process-for-reviewing-and-approving-third-party-service-providers + - page: Please describe the access review process for AssemblyAI internal systems. + path: pages/faq/privacy-security/please-describe-the-access-review-process-for-assemblyai-internal-systems.mdx + slug: /please-describe-the-access-review-process-for-assemblyai-internal-systems + - page: Does AssemblyAI have an incident response plan? + path: pages/faq/privacy-security/please-describe-the-incident-response-plan.mdx + slug: /please-describe-the-incident-response-plan + - page: What are your recovery time and recovery point objectives? + path: pages/faq/privacy-security/what-are-your-recovery-time-and-recovery-point-objectives.mdx + slug: /what-are-your-recovery-time-and-recovery-point-objectives + - page: What does AssemblyAI's cyber insurance cover? + path: pages/faq/privacy-security/what-does-assemblyai-s-cyber-insurance-cover.mdx + slug: /what-does-assemblyai-s-cyber-insurance-cover + - page: What is covered in your security program? + path: pages/faq/privacy-security/what-is-covered-in-your-security-program.mdx + slug: /what-is-covered-in-your-security-program + - page: What is your data retention policy? + path: pages/faq/privacy-security/what-is-your-data-retention-policy.mdx + slug: /what-is-your-data-retention-policy + - page: What is your SLA for repairing Critical/High/Medium vulnerabilities? + path: pages/faq/privacy-security/what-is-your-sla-for-repairing-critical-high-medium-vulnerabilities.mdx + slug: /what-is-your-sla-for-repairing-critical-high-medium-vulnerabilities + - page: What logs are available to customers? + path: pages/faq/privacy-security/what-logs-are-available-to-customers.mdx + slug: /what-logs-are-available-to-customers + - page: What measures are in place for data anonymization or pseudonymization? + path: pages/faq/privacy-security/what-measures-are-in-place-for-data-anonymization-or-pseudonymization.mdx + slug: /what-measures-are-in-place-for-data-anonymization-or-pseudonymization + - page: What TLS versions are supported? + path: pages/faq/privacy-security/what-ssl-tls-versions-and-cipher-suites-does-assemblyai-support.mdx + slug: /what-ssl-tls-versions-and-cipher-suites-does-assemblyai-support + - page: What standards do your internal password policies follow? + path: pages/faq/privacy-security/what-standards-do-your-internal-password-policies-follow.mdx + slug: /what-standards-do-your-internal-password-policies-follow + - page: When will you notify customers if you are breached? + path: pages/faq/privacy-security/when-will-you-notify-customers-if-you-are-breached.mdx + slug: /when-will-you-notify-customers-if-you-are-breached + - page: Where are your servers located? + path: pages/faq/privacy-security/where-are-your-servers-located.mdx + slug: /where-are-your-servers-located + - page: Where do production and backups reside? + path: pages/faq/privacy-security/where-do-production-and-backups-reside.mdx + slug: /where-do-production-and-backups-reside + - section: Pre-recorded audio + skip-slug: true + contents: + - page: Am I charged for transcribing silent audio? + path: pages/faq/async/am-i-charged-for-transcribing-silent-audio.mdx + slug: /am-i-charged-for-transcribing-silent-audio + - page: Are Custom Models More Accurate than General Models? + path: pages/faq/async/are-custom-models-more-accurate-than-general-models.mdx + slug: /are-custom-models-more-accurate-than-general-models + - page: Do I Get Charged for Failed API Calls? + path: pages/faq/async/are-customers-charged-for-api-calls-that-result-in-errors.mdx + slug: /are-customers-charged-for-api-calls-that-result-in-errors + - page: Are there any limits on file size or file duration for files submitted to the API? + path: pages/faq/async/are-there-any-limits-on-file-size-or-file-duration-for-files-submitted-to-the-api.mdx + slug: /are-there-any-limits-on-file-size-or-file-duration-for-files-submitted-to-the-api + - page: Can I customize how words are spelled by the model? + path: pages/faq/async/can-i-customize-how-words-are-spelled-by-the-model.mdx + slug: /can-i-customize-how-words-are-spelled-by-the-model + - page: Can I delete the transcripts I have created using the API? + path: pages/faq/async/can-i-delete-the-transcripts-i-have-created-using-the-api.mdx + slug: /can-i-delete-the-transcripts-i-have-created-using-the-api + - page: Can I get a list of all transcripts I have created? + path: pages/faq/async/can-i-get-a-list-of-all-transcripts-i-have-created.mdx + slug: /can-i-get-a-list-of-all-transcripts-i-have-created + - page: Can I send audio to AssemblyAI in segments and still get speaker labels for the whole recording? + path: pages/faq/async/can-i-send-audio-to-assemblyai-in-segments-and-still-get-speaker-labels-for-the-whole-recording.mdx + slug: /can-i-send-audio-to-assemblyai-in-segments-and-still-get-speaker-labels-for-the-whole-recording + - page: Can I submit files to the API that are stored in a Google Drive? + path: pages/faq/async/can-i-submit-files-to-the-api-that-are-stored-in-a-google-drive.mdx + slug: /can-i-submit-files-to-the-api-that-are-stored-in-a-google-drive + - page: Can I use the API without internet access? + path: pages/faq/async/can-i-use-the-api-without-internet-access.mdx + slug: /can-i-use-the-api-without-internet-access + - page: Do we have resources for building with Make? + path: pages/faq/async/do-we-have-resources-for-building-with-make-.mdx + slug: /do-we-have-resources-for-building-with-make- + - page: Do you have any examples for how to use your API? + path: pages/faq/async/do-you-have-any-examples-for-how-to-use-your-api-.mdx + slug: /do-you-have-any-examples-for-how-to-use-your-api- + - page: Do you have example use cases for using AssemblyAI? + path: pages/faq/async/do-you-have-example-use-cases-for-using-assemblyai-.mdx + slug: /do-you-have-example-use-cases-for-using-assemblyai- + - page: Do you offer cross-file Speaker Identification? + path: pages/faq/async/do-you-offer-cross-file-speaker-identification.mdx + slug: /do-you-offer-cross-file-speaker-identification + - page: Do you offer translation? + path: pages/faq/async/do-you-offer-translation.mdx + slug: /do-you-offer-translation + - page: Do you offer voice-to-voice or text-to-speech (TTS)? + path: pages/faq/async/do-you-offer-voice-to-voice-or-text-to-speech-tts.mdx + slug: /do-you-offer-voice-to-voice-or-text-to-speech-tts + - page: Does it cost extra to export SRT or VTT captions? + path: pages/faq/async/does-it-cost-extra-to-export-srt-or-vtt-captions.mdx + slug: /does-it-cost-extra-to-export-srt-or-vtt-captions + - page: Does it cost more to transcribe an audio or video? + path: pages/faq/async/does-it-cost-more-to-transcribe-an-audio-or-video.mdx + slug: /does-it-cost-more-to-transcribe-an-audio-or-video + - page: Does your API return timestamps for individual words? + path: pages/faq/async/does-your-api-return-timestamps-for-individual-words.mdx + slug: /does-your-api-return-timestamps-for-individual-words + - page: How are individual speakers identified and how does the Speaker Label feature work? + path: pages/faq/async/how-are-individual-speakers-identified-and-how-does-the-speaker-label-feature-work.mdx + slug: /how-are-individual-speakers-identified-and-how-does-the-speaker-label-feature-work + - page: How are paragraphs created for the /paragraphs endpoint? + path: pages/faq/async/how-are-paragraphs-created-for-the-paragraphs-endpoint.mdx + slug: /how-are-paragraphs-created-for-the-paragraphs-endpoint + - page: How are word/transcript level confidence scores calculated? + path: pages/faq/async/how-are-word-transcript-level-confidence-scores-calculated.mdx + slug: /how-are-word-transcript-level-confidence-scores-calculated + - page: How can I integrate AssemblyAI with other services? + path: pages/faq/async/how-can-i-integrate-assemblyai-with-other-services-.mdx + slug: /how-can-i-integrate-assemblyai-with-other-services- + - page: How can I make certain words more likely to be transcribed? + path: pages/faq/async/how-can-i-make-certain-words-more-likely-to-be-transcribed.mdx + slug: /how-can-i-make-certain-words-more-likely-to-be-transcribed + - page: How can I test AssemblyAI without writing code? + path: pages/faq/async/how-can-i-test-assemblyai-without-writing-code-.mdx + slug: /how-can-i-test-assemblyai-without-writing-code- + - page: How can I transcribe YouTube videos? + path: pages/faq/async/how-can-i-transcribe-youtube-videos.mdx + slug: /how-can-i-transcribe-youtube-videos + - page: How can I use Universal-2? + path: pages/faq/async/how-can-i-use-universal-1.mdx + slug: /how-can-i-use-universal-1 + - page: How do I enable Universal-2? + path: pages/faq/async/how-do-i-enable-universal-2.mdx + slug: /how-do-i-enable-universal-2 + - page: How do I generate subtitles? + path: pages/faq/async/how-do-i-generate-subtitles.mdx + slug: /how-do-i-generate-subtitles + - page: How does AssemblyAI compare to other ASR providers? + path: pages/faq/async/how-does-assemblyai-compare-to-other-asr-providers.mdx + slug: /how-does-assemblyai-compare-to-other-asr-providers + - page: How does Automatic Language Detection work? + path: pages/faq/async/how-does-language-detection-work-for-transcriptions.mdx + slug: /how-does-language-detection-work-for-transcriptions + - page: How does the API handle files that contain spoken audio in multiple languages? + path: pages/faq/async/how-does-the-api-handle-files-that-contain-spoken-audio-in-multiple-languages.mdx + slug: /how-does-the-api-handle-files-that-contain-spoken-audio-in-multiple-languages + - page: How long does it take to transcribe a file? + path: pages/faq/async/how-long-does-it-take-to-transcribe-a-file.mdx + slug: /how-long-does-it-take-to-transcribe-a-file + - page: What should I do if I'm getting an error? + path: pages/faq/async/i-am-getting-an-error-what-should-i-do.mdx + slug: /i-am-getting-an-error-what-should-i-do + - page: Is there a Postman collection for using the API? + path: pages/faq/async/is-there-a-postman-collection-for-using-the-api.mdx + slug: /is-there-a-postman-collection-for-using-the-api + - page: Is there a way for us to send the start time / end time for transcription instead of transcribing the whole length of a call recording? + path: pages/faq/async/is-there-a-way-for-us-to-send-the-start-time-end-time-for-transcription-instead-of-transcribing-the-whole-length-of-a-call-recording.mdx + slug: /is-there-a-way-for-us-to-send-the-start-time-end-time-for-transcription-instead-of-transcribing-the-whole-length-of-a-call-recording + - page: Is there an OpenAPI spec/schema for the API? + path: pages/faq/async/is-there-an-openapi-spec-schema-for-the-api.mdx + slug: /is-there-an-openapi-spec-schema-for-the-api + - page: read operation timed out" error + path: pages/faq/async/read-operation-timed-out-error.mdx + slug: /read-operation-timed-out-error + - page: Should I use Speaker Labels or Multi-channel? + path: pages/faq/async/should-i-use-speaker-labels-or-multi-channel.mdx + slug: /should-i-use-speaker-labels-or-multi-channel + - page: Should I use the NA or EU endpoint for my Speech-to-Text requests? + path: pages/faq/async/should-i-use-the-na-or-eu-endpoint.mdx + slug: /should-i-use-the-na-or-eu-endpoint + - page: What are the recommended options for audio noise reduction? + path: pages/faq/async/what-are-the-recommended-options-for-audio-noise-reduction.mdx + slug: /what-are-the-recommended-options-for-audio-noise-reduction + - page: What audio and video file types are supported by your API? + path: pages/faq/async/what-audio-and-video-file-types-are-supported-by-your-api.mdx + slug: /what-audio-and-video-file-types-are-supported-by-your-api + - page: What IP Address Should I Whitelist for AssemblyAI? + path: pages/faq/async/what-ip-address-should-i-whitelist-for-assemblyai.mdx + slug: /what-ip-address-should-i-whitelist-for-assemblyai + - page: What is the difference between Custom Vocabulary and Custom Spelling? + path: pages/faq/async/what-is-the-difference-between-custom-vocabulary-and-custom-spelling.mdx + slug: /what-is-the-difference-between-custom-vocabulary-and-custom-spelling + - page: What is the difference between the Slam-1, Universal, and Nano models? + path: pages/faq/async/what-is-the-difference-between-the-best-and-nano-tiers.mdx + slug: /what-is-the-difference-between-the-best-and-nano-tiers + - page: What is the minimum audio duration that the API can transcribe? + path: pages/faq/async/what-is-the-minimum-audio-duration-that-the-api-can-transcribe.mdx + slug: /what-is-the-minimum-audio-duration-that-the-api-can-transcribe + - page: What is the recommended file type for using your API? + path: pages/faq/async/what-is-the-recommended-file-type-for-using-your-api.mdx + slug: /what-is-the-recommended-file-type-for-using-your-api + - page: What is Universal-2? + path: pages/faq/async/what-is-universal-2.mdx + slug: /what-is-universal-2 + - page: What languages do you support? + path: pages/faq/async/what-languages-do-you-support-.mdx + slug: /what-languages-do-you-support- + - page: What types of audio URLs can I use with the API? + path: pages/faq/async/what-types-of-audio-urls-can-i-use-with-the-api.mdx + slug: /what-types-of-audio-urls-can-i-use-with-the-api + - page: Where can I find a list of recent changes to the API? + path: pages/faq/async/where-can-i-find-a-list-of-recent-changes-to-the-api.mdx + slug: /where-can-i-find-a-list-of-recent-changes-to-the-api + - section: Streaming audio + skip-slug: true + contents: + - page: Can I use speaker diarization with Streaming Speech-to-Text? + path: pages/faq/streaming/can-i-use-speaker-diarization-with-live-audio-transcription.mdx + slug: /can-i-use-speaker-diarization-with-live-audio-transcription + - page: How accurate is your Streaming transcription compared to Async transcription? + path: pages/faq/streaming/how-accurate-is-your-real-time-transcription-compared-to-async-transcription.mdx + slug: /how-accurate-is-your-real-time-transcription-compared-to-async-transcription + - page: How does automatically scaling concurrency for Streaming STT work? + path: pages/faq/streaming/how-does-automatically-scaling-concurrency-for-streaming-stt-work.mdx + slug: /how-does-automatically-scaling-concurrency-for-streaming-stt-work + - page: What languages are supported for Streaming Speech-to-text? + path: pages/faq/streaming/language-support-for-real-time-transcription.mdx + slug: /language-support-for-real-time-transcription + - page: Resolving SSL Certificate Verification Error When Trying to Use Streaming STT + path: pages/faq/streaming/resolving-ssl-certificate-verification-error-in-assemblyai-real-time-transcriber.mdx + slug: /resolving-ssl-certificate-verification-error-in-assemblyai-real-time-transcriber + - page: I am getting a "Model deprecated. See docs for new model information" error message. What does it mean? + path: pages/faq/streaming/upgrading-to-the-universal-streaming-model.mdx + slug: /upgrading-to-the-universal-streaming-model + - page: Why are my Streaming Transcription results repeating? + path: pages/faq/streaming/why-are-my-real-time-transcription-results-repeating.mdx + slug: /why-are-my-real-time-transcription-results-repeating + - section: Speech understanding + skip-slug: true + contents: + - page: How do Content Moderation severity scores work? + path: pages/faq/speech-understanding/how-do-content-moderation-severity-scores-work.mdx + slug: /how-do-content-moderation-severity-scores-work + - page: How can I summarize my audio file? + path: pages/faq/speech-understanding/how-do-your-summarization-models-work.mdx + slug: /how-do-your-summarization-models-work + - page: Is Mistral still supported? + path: pages/faq/speech-understanding/is-mistral-still-supported.mdx + slug: /is-mistral-still-supported + - page: Is pricing for Audio Intelligence per feature or all inclusive? + path: pages/faq/speech-understanding/is-pricing-for-audio-intelligence-per-feature-or-all-inclusive.mdx + slug: /is-pricing-for-audio-intelligence-per-feature-or-all-inclusive + - page: Understanding Input and Output Tokens for LeMUR + path: pages/faq/speech-understanding/understanding-input-and-output-tokens-in-ai-models.mdx + slug: /understanding-input-and-output-tokens-in-ai-models + - section: Playground + skip-slug: true + contents: + - page: Can you use the Playground with files in languages other than English? + path: pages/faq/playground/can-you-use-the-playground-with-files-in-languages-other-than-english.mdx + slug: /can-you-use-the-playground-with-files-in-languages-other-than-english + - page: How do I delete a transcript I created using the Playground? + path: pages/faq/playground/how-do-i-delete-a-transcript-i-created-using-the-playground.mdx + slug: /how-do-i-delete-a-transcript-i-created-using-the-playground + - page: Why is the transcription I am receiving using the Playground in a different language? + path: pages/faq/playground/why-is-the-transcription-i-am-receiving-using-the-playground-in-a-different-language.mdx + slug: /why-is-the-transcription-i-am-receiving-using-the-playground-in-a-different-language + - section: Company + skip-slug: true + contents: + - page: Do you have an affiliate marketing program? + path: pages/faq/company/do-you-have-an-affiliate-marketing-program.mdx + slug: /do-you-have-an-affiliate-marketing-program + - page: Do you have any job openings or internship opportunities? + path: pages/faq/company/do-you-have-any-job-openings-or-internship-opportunities.mdx + slug: /do-you-have-any-job-openings-or-internship-opportunities + - page: How do I contact support? + path: pages/faq/company/how-do-i-contact-support.mdx + slug: /how-do-i-contact-support + - page: How do I get in touch with your Sales team? + path: pages/faq/company/how-do-i-get-in-touch-with-your-sales-team.mdx + slug: /how-do-i-get-in-touch-with-your-sales-team + - page: I’ve spotted an issue with the website, what should I do? + path: pages/faq/company/ive-spotted-an-issue-with-the-website-what-should-i-do.mdx + slug: /ive-spotted-an-issue-with-the-website-what-should-i-do + - page: What are your support hours and response time SLAs? + path: pages/faq/company/what-are-your-support-hours-and-response-time-slas.mdx + slug: /what-are-your-support-hours-and-response-time-slas - tab: playground - tab: changelog colors: diff --git a/fern/pages/faq/accounts-billing-pricing/can-an-account-have-more-than-one-api-key.mdx b/fern/pages/faq/accounts-billing-pricing/can-an-account-have-more-than-one-api-key.mdx new file mode 100644 index 00000000..cfac9de4 --- /dev/null +++ b/fern/pages/faq/accounts-billing-pricing/can-an-account-have-more-than-one-api-key.mdx @@ -0,0 +1,5 @@ +--- +title: "Can an account have more than one API key?" +--- + +Yes, accounts can have multiple API keys! You can create additional API keys in your [dashboard](https://www.assemblyai.com/app/api-keys). You can also organize those API keys within "Projects". See [our official documentation](https://www.assemblyai.com/docs/deployment/account-management) to learn more. diff --git a/fern/pages/faq/accounts-billing-pricing/can-i-sign-up-for-free.mdx b/fern/pages/faq/accounts-billing-pricing/can-i-sign-up-for-free.mdx new file mode 100644 index 00000000..17fb9faa --- /dev/null +++ b/fern/pages/faq/accounts-billing-pricing/can-i-sign-up-for-free.mdx @@ -0,0 +1,16 @@ +--- +title: "Can I sign up for free?" +--- + +Yes! You can use the API for free on a limited trial. Simply add a credit card to your account at any time to upgrade from the trial. + +An account on the free tier has access to our pre-recorded audio transcription and Audio Intelligence features using our Async `/v2/transcript` endpoint. It also has access to our streaming STT feature using our `wss://streaming.assemblyai.com/v3/ws` endpoint. + +An account on the free tier can transcribe up to \$50 in transcription and will be able to process up to five files concurrently. If you upgrade your account before using \$50 in transcription any unused amount will be retained on your account. + +Once you have transcribed \$50 worth of files you will need to upgrade your account to continue using the API. + + + The free tier does not include access to LeMUR. To use LeMUR, you will need to + upgrade your account by adding a credit card. + diff --git a/fern/pages/faq/accounts-billing-pricing/do-my-free-credits-expire.mdx b/fern/pages/faq/accounts-billing-pricing/do-my-free-credits-expire.mdx new file mode 100644 index 00000000..467a710e --- /dev/null +++ b/fern/pages/faq/accounts-billing-pricing/do-my-free-credits-expire.mdx @@ -0,0 +1,5 @@ +--- +title: "Do my free credits expire?" +--- + +No - your free credits don't expire and will remain in your account until used, even if you upgrade your account. diff --git a/fern/pages/faq/accounts-billing-pricing/do-you-have-any-type-of-special-pricing-for-non-profit-companies.mdx b/fern/pages/faq/accounts-billing-pricing/do-you-have-any-type-of-special-pricing-for-non-profit-companies.mdx new file mode 100644 index 00000000..9dd081ae --- /dev/null +++ b/fern/pages/faq/accounts-billing-pricing/do-you-have-any-type-of-special-pricing-for-non-profit-companies.mdx @@ -0,0 +1,5 @@ +--- +title: "Do you have any type of special pricing for non-profit companies?" +--- + +In some cases we are able to offer discounts to non-profit companies. To learn more please reach out to [support@assemblyai.com](mailto:support@assemblyai.com) with more information about your company and use case and we would happy to discuss how we might be able to help. diff --git a/fern/pages/faq/accounts-billing-pricing/do-you-offer-a-student-discount.mdx b/fern/pages/faq/accounts-billing-pricing/do-you-offer-a-student-discount.mdx new file mode 100644 index 00000000..5583c7e1 --- /dev/null +++ b/fern/pages/faq/accounts-billing-pricing/do-you-offer-a-student-discount.mdx @@ -0,0 +1,5 @@ +--- +title: "Do you offer a student discount?" +--- + +We do offer student discounts for accounts associated with valid educational email addresses. To learn more please reach out to [support@assemblyai.com](mailto:support@assemblyai.com) with the email address associated with your account and information on your educational institution and we would be happy to help! diff --git a/fern/pages/faq/accounts-billing-pricing/do-you-offer-special-pricing-for-start-ups.mdx b/fern/pages/faq/accounts-billing-pricing/do-you-offer-special-pricing-for-start-ups.mdx new file mode 100644 index 00000000..803aeabc --- /dev/null +++ b/fern/pages/faq/accounts-billing-pricing/do-you-offer-special-pricing-for-start-ups.mdx @@ -0,0 +1,5 @@ +--- +title: "Do you offer special pricing for start-ups?" +--- + +Startups can apply for our AssemblyAI Startup Program which is designed for early-stage startups to innovate and build powerful products from voice data without financial constraints, using the industry-leading speech-to-text API. [Apply here](https://www.assemblyai.com/startup-program). diff --git a/fern/pages/faq/accounts-billing-pricing/do-you-offer-special-pricing-for-ycombinator-companies.mdx b/fern/pages/faq/accounts-billing-pricing/do-you-offer-special-pricing-for-ycombinator-companies.mdx new file mode 100644 index 00000000..b1b386b6 --- /dev/null +++ b/fern/pages/faq/accounts-billing-pricing/do-you-offer-special-pricing-for-ycombinator-companies.mdx @@ -0,0 +1,5 @@ +--- +title: "Do you offer special pricing for YCombinator companies?" +--- + +We do offer special pricing for YCombinator companies! Please reach out to us [here](https://www.assemblyai.com/contact) to discuss this discount and see if you qualify. diff --git a/fern/pages/faq/accounts-billing-pricing/do-you-offer-volume-discounts.mdx b/fern/pages/faq/accounts-billing-pricing/do-you-offer-volume-discounts.mdx new file mode 100644 index 00000000..c73db158 --- /dev/null +++ b/fern/pages/faq/accounts-billing-pricing/do-you-offer-volume-discounts.mdx @@ -0,0 +1,5 @@ +--- +title: "Do you offer volume discounts?" +--- + +Absolutely! If you plan to send large volumes of audio and video content through our API, please reach out to us [here](https://www.assemblyai.com/contact) to see if you qualify for a volume discount. diff --git a/fern/pages/faq/accounts-billing-pricing/how-do-i-add-developers-to-my-team.mdx b/fern/pages/faq/accounts-billing-pricing/how-do-i-add-developers-to-my-team.mdx new file mode 100644 index 00000000..4bd53b27 --- /dev/null +++ b/fern/pages/faq/accounts-billing-pricing/how-do-i-add-developers-to-my-team.mdx @@ -0,0 +1,32 @@ +--- +title: "How do I add developers to my account?" +--- + +Currently, our system supports only one email account per user account. This means that adding multiple developers to a single team account is not directly possible at this time. + +## Recommended workaround + +To share access among team members, we recommend using a password manager like 1Password. This allows you to securely share the account credentials between multiple developers. + +## Steps to share access: + +1. Set up a shared vault in your password manager +2. Store the account credentials in this shared vault +3. Invite team members to access the shared vault +4. Team members can now use the shared credentials to access the account + + + When sharing credentials, ensure all team members follow best practices for + security, such as using strong, unique passwords and enabling two-factor + authentication where possible. + + +## Future updates + +We are actively working on improving our dashboard to support multi-user accounts. These improvements are expected to be available towards the end of Q4. Once released, you'll be able to add multiple developers to your team directly within the platform. + + + If you are not already, consider following [our + Changelog](https://www.assemblyai.com/changelog) to learn about new features + and improvements as they are released. + diff --git a/fern/pages/faq/accounts-billing-pricing/how-do-i-change-my-account-email-address.mdx b/fern/pages/faq/accounts-billing-pricing/how-do-i-change-my-account-email-address.mdx new file mode 100644 index 00000000..5b428c67 --- /dev/null +++ b/fern/pages/faq/accounts-billing-pricing/how-do-i-change-my-account-email-address.mdx @@ -0,0 +1,5 @@ +--- +title: "How do I change my account email address?" +--- + +To change your email address, send a request to [support@assemblyai.com](mailto:support@assemblyai.com) from your current registered email address, specifying the new email you'd like to use. diff --git a/fern/pages/faq/accounts-billing-pricing/how-do-i-update-company-information-on-my-invoices.mdx b/fern/pages/faq/accounts-billing-pricing/how-do-i-update-company-information-on-my-invoices.mdx new file mode 100644 index 00000000..33b0d717 --- /dev/null +++ b/fern/pages/faq/accounts-billing-pricing/how-do-i-update-company-information-on-my-invoices.mdx @@ -0,0 +1,5 @@ +--- +title: "How do I update company information on my invoices?" +--- + +To update your billing information, please email [support@assemblyai.com](mailto:support@assemblyai.com) with the details you would like to appear on your invoices. diff --git a/fern/pages/faq/accounts-billing-pricing/how-does-auto-pay-work.mdx b/fern/pages/faq/accounts-billing-pricing/how-does-auto-pay-work.mdx new file mode 100644 index 00000000..570efdda --- /dev/null +++ b/fern/pages/faq/accounts-billing-pricing/how-does-auto-pay-work.mdx @@ -0,0 +1,27 @@ +--- +title: "How does Auto-pay work?" +--- + +Auto-pay automatically recharges your account when your balance falls below a specified threshold. When triggered, it charges your card to bring your balance back to a predetermined amount. Auto-pay is recommended for production environments to prevent service interruptions. + + + + + +### Example + +If you set: + +- Whenever my balance falls below: \$25 +- Bring my balance back to: \$50 + +When your balance of \$26 drops to \$21 after a \$5 charge, auto-pay will add \$29 to reach your \$50 target balance. + + + When your balance hits \$0 without auto-pay, your API access is restricted + until you add funds. + + +Enable auto-pay and maintain a healthy balance to ensure uninterrupted API access. Set your threshold based on your typical monthly usage to avoid frequent small charges. + +You can also disable auto-pay at any time through your account settings. diff --git a/fern/pages/faq/accounts-billing-pricing/how-does-pricing-work.mdx b/fern/pages/faq/accounts-billing-pricing/how-does-pricing-work.mdx new file mode 100644 index 00000000..c4069100 --- /dev/null +++ b/fern/pages/faq/accounts-billing-pricing/how-does-pricing-work.mdx @@ -0,0 +1,30 @@ +--- +title: "How does pricing work?" +--- + +Our pricing is Pay-As-You-Go. For pre-recorded audio, pricing is based on the duration of the audio file submitted, the Speech Model selected, and the features used in the request. For streaming audio, pricing is based on the duration of the streaming session. + +Current pricing for pre-recorded audio on the Universal tier is \$0.27 per hour, and the Nano tier is \$0.12 per hour. Streaming STT is \$0.15 per hour. Audio Intelligence pricing depends on the models being used. LeMUR pricing is based on input and output size. + +You can find the breakdown of all our pricing information on our [Pricing page](https://www.assemblyai.com/pricing)! + +## Pricing Structure (pre-recorded audio) + +- We charge based on the actual duration of your audio/video files in seconds +- Rates are listed per hour for simplicity (e.g., \$0.27/hour), but we pro-rate to the exact second +- Additional features (like Entity Detection) are also charged at their respective hourly rates + +## Pricing Structure (streaming audio) + +- We charge based on the duration of the created session + +## Credit System + +- Add credits to your account balance +- Credits are deducted only after successful transcription completion +- Failed transcripts are not charged +- We don't offer unlimited usage plans or monthly subscriptions - we only charge you for what you use. + +## Example + +A 30-minute file at \$0.12/hour would cost \$0.06 (30 minutes = 0.5 hours × \$0.12). If you add Entity Detection at \$0.08/hour, the total would be \$0.10 `(\$0.12 + \$0.08) × 0.5`. diff --git a/fern/pages/faq/accounts-billing-pricing/how-does-the-concurrency-limit-work-for-transcription-requests.mdx b/fern/pages/faq/accounts-billing-pricing/how-does-the-concurrency-limit-work-for-transcription-requests.mdx new file mode 100644 index 00000000..5c47788d --- /dev/null +++ b/fern/pages/faq/accounts-billing-pricing/how-does-the-concurrency-limit-work-for-transcription-requests.mdx @@ -0,0 +1,48 @@ +--- +title: "How does the concurrency limit work for transcription requests?" +--- + +AssemblyAI's transcription service for pre-recorded audio uses a concurrency limit to manage the number of simultaneous transcription jobs that can be processed for each account. This article explains how the concurrency limit works and what happens when you reach it. + +## Understanding the Concurrency Limit + +The concurrency limit determines how many transcription jobs can be processed simultaneously for your account. When you submit transcription requests: + +- If you're below the concurrency limit, your job starts processing immediately. +- If you've reached the limit, additional jobs are placed in a queue. + +## Queue Processing + +When you've reached your concurrency limit: + +1. New transcription requests are added to a queue. +2. As soon as a currently processing job finishes, the next job in the queue starts processing. +3. This continues until all queued jobs are processed. + +The time it takes for queued jobs to start processing depends on the duration of the audio files currently being transcribed. + +## Increasing Your Concurrency Limit + +If you find that the current concurrency limit is insufficient for your needs, you can increase it by upgrading your account. To do this: + +1. Add a credit card to your account. +2. Choose a plan that offers a higher concurrency limit. + +Upgrading your account not only increases your concurrency limit but may also provide additional features and benefits. + + + We offer custom concurrency limits that scale to support any workload at no + additional cost. If you need a higher concurrency limit please either [contact + our Sales team](https://www.assemblyai.com/contact) or reach out to us at + [support@assemblyai.com](mailto:support@assemblyai.com). + + +## Best Practices + +To optimize your use of the transcription service: + +- Monitor your job queue to anticipate processing times. +- For time-sensitive transcriptions, consider upgrading your account if you frequently hit the concurrency limit. +- Batch your transcription requests efficiently to make the most of your current limit. + +If you have any questions about your concurrency limit or need assistance with upgrading your account, please contact our support team. diff --git a/fern/pages/faq/accounts-billing-pricing/how-often-does-the-usage-and-spend-information-in-my-dashboard-update-.mdx b/fern/pages/faq/accounts-billing-pricing/how-often-does-the-usage-and-spend-information-in-my-dashboard-update-.mdx new file mode 100644 index 00000000..b17708a9 --- /dev/null +++ b/fern/pages/faq/accounts-billing-pricing/how-often-does-the-usage-and-spend-information-in-my-dashboard-update-.mdx @@ -0,0 +1,16 @@ +--- +title: "How often does the Usage and Spend Information in my Dashboard update?" +--- + +You can find your Usage and Spending information tracked [in your Dashboard](https://www.assemblyai.com/dashboard/cost) as you transcribe using the API. + +Your current account balance can be found here here: [https://www.assemblyai.com/dashboard/account/billing](https://www.assemblyai.com/dashboard/account/billing) + +Your transcription history can be found here: [https://www.assemblyai.com/dashboard/transcription-history](https://www.assemblyai.com/dashboard/transcription-history) + +The usage and spending information in the Dashboard is updated every 2 minutes. + + + Real-Time transcription sessions must be closed for the information to + populate. + diff --git a/fern/pages/faq/accounts-billing-pricing/how-to-delete-your-account.mdx b/fern/pages/faq/accounts-billing-pricing/how-to-delete-your-account.mdx new file mode 100644 index 00000000..993b6028 --- /dev/null +++ b/fern/pages/faq/accounts-billing-pricing/how-to-delete-your-account.mdx @@ -0,0 +1,18 @@ +--- +title: "How to Delete Your Account" +--- + +If you need to delete your account and all associated data, you can do so easily through your Dashboard. Follow these steps: + +1. Log in to your account +2. Navigate to the [Account page](https://www.assemblyai.com/dashboard/account/personal) on your Dashboard +3. Look for the "Delete Account" button +4. Follow the prompts to complete the account deletion process + +## Important notes + +- Account deletion is permanent and cannot be undone +- All data associated with your account will be removed +- Make sure to back up any important information before proceeding with deletion + +If you encounter any issues during the deletion process or have additional questions, please contact our support team for assistance. diff --git a/fern/pages/faq/accounts-billing-pricing/how-to-get-your-api-key.mdx b/fern/pages/faq/accounts-billing-pricing/how-to-get-your-api-key.mdx new file mode 100644 index 00000000..c0b71cb8 --- /dev/null +++ b/fern/pages/faq/accounts-billing-pricing/how-to-get-your-api-key.mdx @@ -0,0 +1,23 @@ +--- +title: "How to an get AssemblyAI API Key" +--- + +To obtain your API key for AssemblyAI, follow these simple steps: + +1. Visit the [AssemblyAI login page](https://www.assemblyai.com/dashboard/login). + +2. Log into your account. If you don't have an account yet, you'll need to sign up first. + +3. Once logged in, navigate to your [API Keys page](https://www.assemblyai.com/dashboard/api-keys). + +4. Your API key will be displayed here: + + + + + +## Notes + +- Keep your API key secure and do not share it publicly. +- If you need to refresh your API key for any reason, you can do so from the same dashboard. +- Your API key is required to authenticate your requests when using the AssemblyAI API. diff --git a/fern/pages/faq/accounts-billing-pricing/how-to-share-account-access-with-team-members.mdx b/fern/pages/faq/accounts-billing-pricing/how-to-share-account-access-with-team-members.mdx new file mode 100644 index 00000000..0a540108 --- /dev/null +++ b/fern/pages/faq/accounts-billing-pricing/how-to-share-account-access-with-team-members.mdx @@ -0,0 +1,33 @@ +--- +title: "How to Share Account Access with Team Members" +--- + +While our platform only allows a single email address and API key per account, there are ways to effectively share access with your team members. Here's how you can manage shared access: + +## Creating a Shared Team Account + +The recommended approach for team access is to create a shared account: + +1. Set up a new account using a shared email address (e.g., dev@yourcompany.com) +2. Securely share the account credentials with team members using a password manager + +## Best Practices for Shared Accounts + +- Use a generic email address that isn't tied to a specific individual +- Implement strong password policies +- Regularly review and update access as team members join or leave + +## Changing the Email Address on an Existing Account + +If you need to update the email address associated with your account to a shared email: + +1. Contact our support team +2. Request an email address change for your account +3. Provide the new shared email address you'd like to use + + + Remember: Each account can only have one email address and API key. Changing + the email will affect all access associated with the account. + + +By following these guidelines, you can effectively manage shared access to your account while maintaining security and control over your team's usage of our platform. diff --git a/fern/pages/faq/accounts-billing-pricing/is-assemblyai-available-on-the-aws-marketplace.mdx b/fern/pages/faq/accounts-billing-pricing/is-assemblyai-available-on-the-aws-marketplace.mdx new file mode 100644 index 00000000..9b3168ca --- /dev/null +++ b/fern/pages/faq/accounts-billing-pricing/is-assemblyai-available-on-the-aws-marketplace.mdx @@ -0,0 +1,5 @@ +--- +title: "Is AssemblyAI available on the AWS Marketplace?" +--- + +We are! We can be found on the Marketplace [here](https://aws.amazon.com/marketplace/pp/prodview-j45hta6jdej7c?sr=0-1&ref_=beagle&applicationId=AWSMPContessa). Please [connect with our Sales team](https://www.assemblyai.com/contact/sales) to learn more. diff --git a/fern/pages/faq/accounts-billing-pricing/what-are-my-concurrency-limits.mdx b/fern/pages/faq/accounts-billing-pricing/what-are-my-concurrency-limits.mdx new file mode 100644 index 00000000..5ed734a7 --- /dev/null +++ b/fern/pages/faq/accounts-billing-pricing/what-are-my-concurrency-limits.mdx @@ -0,0 +1,39 @@ +--- +title: "What are my concurrency limits?" +--- + +A concurrency limit is the number of requests that can be processed for an account at any given time. + +Each account has an assigned concurrency limit (also referred to as a throttle limit). For free accounts, the default is 5 for Asynchronous and Streaming transcriptions. For upgraded accounts, the default is 200 for Asynchronous transcriptions and 100 for Streaming transcriptions. + +You can check your concurrency limits [on your dashboard](https://www.assemblyai.com/dashboard/rate-limits). + +### Asynchronous speech-to-text limits + +Below are the default limits for how many requests you can have processing in parallel when submitting jobs to our `/v2/transcript` endpoint. + +| Account type | Concurrency limit | +| ------------ | ----------------- | +| Free | 5 | +| Pro | 200 | + +### Streaming speech-to-text limits + +Below are the default limits for how many Streaming sessions you can have open in parallel. + +| Account type | Concurrency limit | +| ------------ | ----------------- | +| Free | 5 | +| Pro | 100 | + +Note: Our Streaming STT feature includes automatically scaling concurrency. Anytime you are using 70% or more of your streaming concurrency, your concurrency limit will automatically increase by 10% every sixty seconds. + +You can find more information on Concurrency Limits in our Documentation [here](https://www.assemblyai.com/docs/concepts/concurrency-limit)! + +### Need a higher concurrency? + +We offer custom concurrency limits that scale to support any workload at no additional cost. If you need a higher concurrency limit please either [contact our Sales team](https://www.assemblyai.com/contact) or reach out to us at [support@assemblyai.com](mailto:support@assemblyai.com). + +### Rate Limits + +In addition to the concurrency limit, there's a rate limit for the API, which restricts users to a maximum of 20,000 requests per five minutes. diff --git a/fern/pages/faq/accounts-billing-pricing/what-happens-if-i-reach-my-concurrency-limit.mdx b/fern/pages/faq/accounts-billing-pricing/what-happens-if-i-reach-my-concurrency-limit.mdx new file mode 100644 index 00000000..aae73fa5 --- /dev/null +++ b/fern/pages/faq/accounts-billing-pricing/what-happens-if-i-reach-my-concurrency-limit.mdx @@ -0,0 +1,22 @@ +--- +title: "What happens if I reach my concurrency limit?" +--- + +### Async speech-to-text + +When you reach your concurrency limit, new requests are automatically queued. As soon as a current transcription completes, the next job in the queue begins processing. All transcripts will be processed, though queued jobs may take longer than usual. + +For example, if a user's concurrency limit is set to 200 and they submit 201 transcript requests, the user will receive a throttle alert email. The first 200 transcription requests will begin to process immediately. The 201st one will wait until one of the previous 200 requests has finished before processing. + +### Streaming + +When you reach your streaming concurrency limit, new requests will receive a `1008` error with the message "`Unauthorized connection: Too many concurrent sessions`". + +Note: Our Streaming STT feature includes automatically scaling concurrency. Anytime you are using 70% or more of your streaming concurrency your concurrency limit will automatically increase by 10% every sixty seconds. + + + We offer custom concurrency limits that scale to support any workload at no + additional cost. If you need a higher concurrency limit, please either + [contact our Sales team](https://www.assemblyai.com/contact/sales) or reach + out to us at [support@assemblyai.com](mailto:support@assemblyai.com). + diff --git a/fern/pages/faq/accounts-billing-pricing/what-payment-methods-do-you-accept.mdx b/fern/pages/faq/accounts-billing-pricing/what-payment-methods-do-you-accept.mdx new file mode 100644 index 00000000..fde140b9 --- /dev/null +++ b/fern/pages/faq/accounts-billing-pricing/what-payment-methods-do-you-accept.mdx @@ -0,0 +1,5 @@ +--- +title: "What payment methods do you accept?" +--- + +We accept all major credit cards. ACH transfers are also available in some cases - please contact [support@assemblyai.com](mailto:support@assemblyai.com) to learn more about this option. diff --git a/fern/pages/faq/async/am-i-charged-for-transcribing-silent-audio.mdx b/fern/pages/faq/async/am-i-charged-for-transcribing-silent-audio.mdx new file mode 100644 index 00000000..8832f5ad --- /dev/null +++ b/fern/pages/faq/async/am-i-charged-for-transcribing-silent-audio.mdx @@ -0,0 +1,7 @@ +--- +title: "Am I charged for transcribing silent audio?" +--- + +Yes, our billing is based on audio duration, not the amount of speech content. However, you can use the `speech_threshold` parameter to avoid processing files with minimal speech content. This parameter lets you specify a minimum percentage of speech required for transcription. + +[Learn more about speech threshold](/docs/speech-to-text/pre-recorded-audio/speech-threshold) diff --git a/fern/pages/faq/async/are-custom-models-more-accurate-than-general-models.mdx b/fern/pages/faq/async/are-custom-models-more-accurate-than-general-models.mdx new file mode 100644 index 00000000..a2532de6 --- /dev/null +++ b/fern/pages/faq/async/are-custom-models-more-accurate-than-general-models.mdx @@ -0,0 +1,11 @@ +--- +title: "Are Custom Models More Accurate than General Models?" +--- + +In the field of ASR, custom models are rarely more accurate than the best general models (learn more about one measure of accuracy, [Word Error Rate or WER](https://www.assemblyai.com/blog/word-error-rate/?utm_source=google&utm_medium=cpc&utm_campaign=brand), here). This is because general models are trained on huge datasets, and are constantly maintained and updated using the latest deep learning research. + +For example, at AssemblyAI, we train large deep neural networks on over 12 million hours of speech data. This training data is a mix of many different types of audio (broadcast TV recordings, phone calls, Zoom meetings, videos, etc), accents, and speakers. This massive amount of diverse training data helps our ASR models to generalize extremely well across all types of audio/data, speakers, recording quality, and accents when converting Speech-to-Text in the real world. + +Custom models usually come into the mix when dealing with audio data that have unique characteristics unseen by a general model. However, because large, accurate general models see most types of audio data during training, there are not many "unique characteristics" that would trip up a general model - or that a custom model would even be able to learn. + +To learn more about this topic, see [this blog post](https://www.assemblyai.com/blog/do-i-need-a-custom-speech-recognition-model/). diff --git a/fern/pages/faq/async/are-customers-charged-for-api-calls-that-result-in-errors.mdx b/fern/pages/faq/async/are-customers-charged-for-api-calls-that-result-in-errors.mdx new file mode 100644 index 00000000..a3cf81a4 --- /dev/null +++ b/fern/pages/faq/async/are-customers-charged-for-api-calls-that-result-in-errors.mdx @@ -0,0 +1,11 @@ +--- +title: "Do I Get Charged for Failed API Calls?" +--- + +You won't be charged for failed API calls. If your transcript response returns an "error" status, the request won't count against your credits. + +You can review any failed transcripts in your [Async Transcription Processing Queue](https://www.assemblyai.com/dashboard/transcription-history). + + + + diff --git a/fern/pages/faq/async/are-there-any-limits-on-file-size-or-file-duration-for-files-submitted-to-the-api.mdx b/fern/pages/faq/async/are-there-any-limits-on-file-size-or-file-duration-for-files-submitted-to-the-api.mdx new file mode 100644 index 00000000..0bdcf562 --- /dev/null +++ b/fern/pages/faq/async/are-there-any-limits-on-file-size-or-file-duration-for-files-submitted-to-the-api.mdx @@ -0,0 +1,7 @@ +--- +title: "Are there any limits on file size or file duration for files submitted to the API?" +--- + +There is both a maximum file size and a maximum audio duration for files that can be submitted to our API for transcription. The maximum file size that can be submitted to the `/v2/transcript` endpoint for transcription is 5GB, and the maximum duration is 10 hours. + +There is also a maximum file size for a local file uploaded to the API via the `/v2/upload` endpoint, which is 2.2GB. diff --git a/fern/pages/faq/async/can-i-customize-how-words-are-spelled-by-the-model.mdx b/fern/pages/faq/async/can-i-customize-how-words-are-spelled-by-the-model.mdx new file mode 100644 index 00000000..ec1d5fde --- /dev/null +++ b/fern/pages/faq/async/can-i-customize-how-words-are-spelled-by-the-model.mdx @@ -0,0 +1,7 @@ +--- +title: "Can I customize how words are spelled by the model?" +--- + +Yes. The Custom Spelling feature gives you the ability to specify how words are spelled or formatted in the transcript text. For example, Custom Spelling could be used to change the spelling of all instances of the word "Ariana" to "Arianna". It could also be used to change the formatting of "CS 50" to "CS50". + +For more information, check our docs on [Custom Spelling](/docs/speech-to-text/pre-recorded-audio/custom-spelling). diff --git a/fern/pages/faq/async/can-i-delete-the-transcripts-i-have-created-using-the-api.mdx b/fern/pages/faq/async/can-i-delete-the-transcripts-i-have-created-using-the-api.mdx new file mode 100644 index 00000000..29c0b918 --- /dev/null +++ b/fern/pages/faq/async/can-i-delete-the-transcripts-i-have-created-using-the-api.mdx @@ -0,0 +1,7 @@ +--- +title: "Can I delete the transcripts I have created using the API?" +--- + +You can! Completed transcripts are stored indefinitely but they can be deleted at any time using the `DELETE` endpoint covered on [this page](https://www.assemblyai.com/docs/speech-to-text/pre-recorded-audio#delete-transcripts) of our documentation. + +If you are looking to delete transcriptions created via our Playground please reach out to our Support team at [support@assemblyai.com](mailto:support@assemblyai.com). diff --git a/fern/pages/faq/async/can-i-get-a-list-of-all-transcripts-i-have-created.mdx b/fern/pages/faq/async/can-i-get-a-list-of-all-transcripts-i-have-created.mdx new file mode 100644 index 00000000..bae3ace6 --- /dev/null +++ b/fern/pages/faq/async/can-i-get-a-list-of-all-transcripts-i-have-created.mdx @@ -0,0 +1,7 @@ +--- +title: "Can I get a list of all transcripts I have created?" +--- + +You can retrieve a list of all transcripts that you have created by [making a `GET` request to the API](https://assemblyai.com/docs/api-reference/transcripts/list). + +Note historical transcripts can only be retrieved for the last 90 days via this endpoint. If you need transcript information from farther back please reach out to our Support team at support@assemblyai.com. diff --git a/fern/pages/faq/async/can-i-send-audio-to-assemblyai-in-segments-and-still-get-speaker-labels-for-the-whole-recording.mdx b/fern/pages/faq/async/can-i-send-audio-to-assemblyai-in-segments-and-still-get-speaker-labels-for-the-whole-recording.mdx new file mode 100644 index 00000000..0d546411 --- /dev/null +++ b/fern/pages/faq/async/can-i-send-audio-to-assemblyai-in-segments-and-still-get-speaker-labels-for-the-whole-recording.mdx @@ -0,0 +1,21 @@ +--- +title: "Can I send audio to AssemblyAI in segments and still get speaker labels for the whole recording?" +--- + +While AssemblyAI doesn't provide a built-in way to send audio in segments and receive speaker labels for the entire recording, you can achieve this functionality using TitaNet. We have a detailed cookbook that demonstrates this workflow. + +## Using TitaNet for Segmented Audio Processing + +To process audio in segments while maintaining speaker labels across the entire recording: + +1. Use TitaNet to segment your audio +2. Send these segments to AssemblyAI for transcription +3. Combine the results to get speaker labels for the whole recording + +## Step-by-Step Guide + +For a detailed walkthrough of this process, including code examples, please refer to our cookbook: + +[Speaker Diarization with Async Chunking](/docs/guides/speaker-diarization-with-async-chunking) + +This cookbook provides a comprehensive guide on how to implement speaker diarization with asynchronous chunking, allowing you to process large audio files in segments while maintaining accurate speaker labels across the entire recording. diff --git a/fern/pages/faq/async/can-i-submit-files-to-the-api-that-are-stored-in-a-google-drive.mdx b/fern/pages/faq/async/can-i-submit-files-to-the-api-that-are-stored-in-a-google-drive.mdx new file mode 100644 index 00000000..2ecc6a44 --- /dev/null +++ b/fern/pages/faq/async/can-i-submit-files-to-the-api-that-are-stored-in-a-google-drive.mdx @@ -0,0 +1,11 @@ +--- +title: "Can I submit files to the API that are stored in a Google Drive?" +--- + +Yes, you can but the file needs to be shared publicly, smaller than 100MB, and the URL needs to be a downloadable link, which should look something like this: + +``` +https://drive.google.com/u/0/uc?id=1OMjFJ1VUX0ES5tMSdslybvPLxYTjdqLZ&export=download +``` + +See [this guide](/docs/guides/transcribing-google-drive-file) from our AssemblyAI Cookbooks for a detailed walkthrough of how to submit files from Google Drive. diff --git a/fern/pages/faq/async/can-i-use-the-api-without-internet-access.mdx b/fern/pages/faq/async/can-i-use-the-api-without-internet-access.mdx new file mode 100644 index 00000000..c922aae3 --- /dev/null +++ b/fern/pages/faq/async/can-i-use-the-api-without-internet-access.mdx @@ -0,0 +1,5 @@ +--- +title: "Can I use the API without internet access?" +--- + +No - you won't be able to reach AssemblyAI's API without an internet connection. A fast and reliable internet connection is especially recommended when using our upload endpoint or Streaming services. diff --git a/fern/pages/faq/async/do-we-have-resources-for-building-with-make-.mdx b/fern/pages/faq/async/do-we-have-resources-for-building-with-make-.mdx new file mode 100644 index 00000000..82dc25c6 --- /dev/null +++ b/fern/pages/faq/async/do-we-have-resources-for-building-with-make-.mdx @@ -0,0 +1,30 @@ +--- +title: "Do we have resources for building with Make?" +--- + +Yes, we have a Make integration which you can find here: [https://www.make.com/en/integrations/assembly-ai](https://www.make.com/en/integrations/assembly-ai) + +Make is a no-code platform that enables users to build their own workflows through their drag-and-drop features. Users can automate tasks involving voice and text data processing using AssemblyAI on the Make platform without needing to invest in custom software development. + +You can also build an application that connects to various other platforms such as Gmail, Slack, Stripe, etc. For example, an audio file is transcribed by AssemblyAI, and the transcription results are then uploaded to a Google Drive. + +The following modules are supported: + +- Upload a File +- Transcripts +- Transcribe an Audio File (submit or poll) +- Wait until Transcript is Ready (explicitly poll after submit) +- Watch for Transcript Ready Notification (this is the transcript webhook) +- Get a Transcript +- Get Paragraphs of a Transcript +- Get Sentences of a Transcript +- Get Subtitles for a Transcript +- Get Redacted Audio of a Transcript (including file download) +- Search for Words in a Transcript +- List Transcripts +- Delete a Transcript +- LeMUR +- Run a Task using LeMUR +- Purge a LeMUR Request +- Other modules +- Make an API Call diff --git a/fern/pages/faq/async/do-you-have-any-examples-for-how-to-use-your-api-.mdx b/fern/pages/faq/async/do-you-have-any-examples-for-how-to-use-your-api-.mdx new file mode 100644 index 00000000..6f5ee894 --- /dev/null +++ b/fern/pages/faq/async/do-you-have-any-examples-for-how-to-use-your-api-.mdx @@ -0,0 +1,31 @@ +--- +title: "Do you have any examples for how to use your API?" +--- + +We have a plethora of resources to help users implement and try out our API! Here are a few resources that could help you. + +## Documentation + +Our documentation provides info on our various features and models, as well as guides on how to use our API that include our code examples: + +[https://www.assemblyai.com/docs/](https://www.assemblyai.com/docs/) + +## Cookbook + +We have use-case-specific cookbooks that may apply to what you're looking to build: + +[https://www.assemblyai.com/docs/guides](https://www.assemblyai.com/docs/guides) + +## YouTube + +We have an active YouTube channel with more tutorials and AI Content. If you prefer video guides, this is the place for you! You can follow our YouTube channel here: + +[https://www.youtube.com/@AssemblyAI](https://www.youtube.com/@AssemblyAI) + +## AssemblyAI Blog + +The AssemblyAI Blog covers product news, company news, the latest AI research, and AssemblyAI tutorials. The tutorials often cover specific integrations and ideas to implement with our API. + +[https://www.assemblyai.com/blog/](https://www.assemblyai.com/blog/) + +For any questions about any of these resources, please reach out to us via email [support@assemblyai.com](mailto:support@assemblyai.com) with a detailed summary of how we can help! diff --git a/fern/pages/faq/async/do-you-have-example-use-cases-for-using-assemblyai-.mdx b/fern/pages/faq/async/do-you-have-example-use-cases-for-using-assemblyai-.mdx new file mode 100644 index 00000000..468ae2a8 --- /dev/null +++ b/fern/pages/faq/async/do-you-have-example-use-cases-for-using-assemblyai-.mdx @@ -0,0 +1,11 @@ +--- +title: "Do you have example use cases for using AssemblyAI?" +--- + +Yes! You can find the AssemblyAI Cookbooks here: + +[https://www.assemblyai.com/docs/guides](https://www.assemblyai.com/docs/guides) + +AssemblyAI Cookbooks are a resource of code examples, guides, and tutorials for using the AssemblyAI API. They are intended to demonstrate and provide an accompanying code example for use cases using our API. Our team is constantly updating this resource based on our user feedback and feature releases. + +Please reach out if you have any ideas or would like to contribute to the code examples! diff --git a/fern/pages/faq/async/do-you-offer-cross-file-speaker-identification.mdx b/fern/pages/faq/async/do-you-offer-cross-file-speaker-identification.mdx new file mode 100644 index 00000000..3b675c19 --- /dev/null +++ b/fern/pages/faq/async/do-you-offer-cross-file-speaker-identification.mdx @@ -0,0 +1,17 @@ +--- +title: "Do you offer cross-file Speaker Identification?" +--- + +Our API currently supports detecting and labeling different speakers within a single audio file (speaker diarization). We currently don't offer native cross-file speaker identification or voice registration. By default, our system outputs speaker labels as "Speaker A," "Speaker B," "Speaker C," etc. Without additional metadata or processing, these labels don't maintain consistency across different recordings. + +However, there are several effective approaches you can implement to achieve speaker identification across multiple recordings. + +One approach is to use LeMUR to match speaker labels with the individuals within the recording. This approach is good for use-cases where the speaker names are spoken in the audio recording. We have a cookbook with detailed instructions on how to do this here: [https://www.assemblyai.com/docs/guides/speaker-identification](https://www.assemblyai.com/docs/guides/speaker-identification) + +For a more sophisticated approach, you can implement speaker identification using audio embeddings. For this you would first submit your audio file to AssemblyAI for diarization with speaker labels, and then use a model like [Nvidia Titanet](https://huggingface.co/nvidia/speakerverification_en_titanet_large) to generate speaker embeddings from the audio. Then, you would match these embeddings against a vector database of known speakers before replacing our generic labels ("Speaker A/B") with actual names. Refer to our speaker identification cookbook here for more details: [https://www.assemblyai.com/docs/guides/titanet-speaker-identification](https://www.assemblyai.com/docs/guides/titanet-speaker-identification) + + + For a deeper understanding on the differences between Speaker Diarization and + Recognition, check out this blog post on the topic: + [https://www.assemblyai.com/blog/speaker-diarization-vs-recognition/](https://www.assemblyai.com/blog/speaker-diarization-vs-recognition/) + diff --git a/fern/pages/faq/async/do-you-offer-translation.mdx b/fern/pages/faq/async/do-you-offer-translation.mdx new file mode 100644 index 00000000..1cdf910e --- /dev/null +++ b/fern/pages/faq/async/do-you-offer-translation.mdx @@ -0,0 +1,12 @@ +--- +title: "Do you offer translation?" +--- + +AssemblyAI does not currently offer translation services. Our API focuses on transcribing speech to text in the original spoken language. + +While we don't provide direct translation, you can find examples of how to translate your transcripts using popular translation services in our Cookbooks. These guides demonstrate post-processing workflows using various commercial and open-source machine translation models. + +**Cookbooks:** + +- [How to translate transcripts](https://www.assemblyai.com/docs/guides/translate_transcripts) +- [Translating an AssemblyAI Subtitle Transcript](https://www.assemblyai.com/docs/guides/translate_subtitles) diff --git a/fern/pages/faq/async/do-you-offer-voice-to-voice-or-text-to-speech-tts.mdx b/fern/pages/faq/async/do-you-offer-voice-to-voice-or-text-to-speech-tts.mdx new file mode 100644 index 00000000..08941541 --- /dev/null +++ b/fern/pages/faq/async/do-you-offer-voice-to-voice-or-text-to-speech-tts.mdx @@ -0,0 +1,5 @@ +--- +title: "Do you offer voice-to-voice or text-to-speech (TTS)?" +--- + +No - AssemblyAI delivers industry-leading speech-to-text and speech understanding only. We don't offer voice-to-voice transformation or text-to-speech services. diff --git a/fern/pages/faq/async/does-it-cost-extra-to-export-srt-or-vtt-captions.mdx b/fern/pages/faq/async/does-it-cost-extra-to-export-srt-or-vtt-captions.mdx new file mode 100644 index 00000000..8cbd40f9 --- /dev/null +++ b/fern/pages/faq/async/does-it-cost-extra-to-export-srt-or-vtt-captions.mdx @@ -0,0 +1,7 @@ +--- +title: "Does it cost extra to export SRT or VTT captions?" +--- + +It does not! + +Customers are not charged an additional amount for exporting `SRT` or `VTT` files. Once a transcription job for a specific transcript ID is completed, the customer will be charged based on the audio duration of the file and the features/models included in that `POST` request. Customers are not charged for `GET` requests including those to export `SRT` or `VTT` captions. diff --git a/fern/pages/faq/async/does-it-cost-more-to-transcribe-an-audio-or-video.mdx b/fern/pages/faq/async/does-it-cost-more-to-transcribe-an-audio-or-video.mdx new file mode 100644 index 00000000..d3ba8c82 --- /dev/null +++ b/fern/pages/faq/async/does-it-cost-more-to-transcribe-an-audio-or-video.mdx @@ -0,0 +1,5 @@ +--- +title: "Does it cost more to transcribe an audio or video?" +--- + +Our pricing is based solely on duration. Audio and video files are charged at the same rate - you only pay for the length of the content, not the file type. diff --git a/fern/pages/faq/async/does-your-api-return-timestamps-for-individual-words.mdx b/fern/pages/faq/async/does-your-api-return-timestamps-for-individual-words.mdx new file mode 100644 index 00000000..d19006c5 --- /dev/null +++ b/fern/pages/faq/async/does-your-api-return-timestamps-for-individual-words.mdx @@ -0,0 +1,9 @@ +--- +title: "Does your API return timestamps for individual words?" +--- + +Yes! The response for a completed request includes `start` and `end` keys. These keys are timestamp values for when a given word, phrase, or sentence starts and ends. These values are in milliseconds and are accurate to within about 400 milliseconds. + +To convert these timestamps from milliseconds to seconds, divide the timestamp value by 1000. + +See [this section](/docs/api-reference/transcripts/get) of our API reference for an example of the JSON response for a completed transcript, which includes these `start` and `end` keys. diff --git a/fern/pages/faq/async/how-are-individual-speakers-identified-and-how-does-the-speaker-label-feature-work.mdx b/fern/pages/faq/async/how-are-individual-speakers-identified-and-how-does-the-speaker-label-feature-work.mdx new file mode 100644 index 00000000..9909aace --- /dev/null +++ b/fern/pages/faq/async/how-are-individual-speakers-identified-and-how-does-the-speaker-label-feature-work.mdx @@ -0,0 +1,9 @@ +--- +title: "How are individual speakers identified and how does the Speaker Label feature work?" +--- + +When an audio file is submitted with `speaker_labels` set to `true` word timings are used to cut the audio into separate chunks of words. Those chunks are fed into a model to build a "speaker embedding", which is a representation of a speaker. An algorithm is then used to cluster speaker embeddings that are similar to each other. If there are two speakers, there will be two distinct clusters of speaker embeddings. Those clusters are then used to assign speakers. + +When the model is determining speakers, the entire audio is taken into consideration. It takes a certain amount of audio for a person to be identified as a unique speaker (we typically say 30 seconds). If a person doesn't speak much over the course of a file or they reply in short phrases like "okay" and "right" they may not get identified as a unique speaker. If they aren't identified as a unique speaker, their words will be attributed to the speaker embedding the model feels is most similar. + +For more information on Speaker Labels, see [this blog post](https://www.assemblyai.com/blog/top-speaker-diarization-libraries-and-apis). diff --git a/fern/pages/faq/async/how-are-paragraphs-created-for-the-paragraphs-endpoint.mdx b/fern/pages/faq/async/how-are-paragraphs-created-for-the-paragraphs-endpoint.mdx new file mode 100644 index 00000000..6ab76e8e --- /dev/null +++ b/fern/pages/faq/async/how-are-paragraphs-created-for-the-paragraphs-endpoint.mdx @@ -0,0 +1,5 @@ +--- +title: "How are paragraphs created for the /paragraphs endpoint?" +--- + +With our `/v2/transcript/{TRANSCRIPT-ID}/paragraphs` endpoint you can export a completed transcript broken into paragraphs. These paragraphs are created using an algorithm that looks at certain heuristics like speaking time, text length, pauses, etc. diff --git a/fern/pages/faq/async/how-are-word-transcript-level-confidence-scores-calculated.mdx b/fern/pages/faq/async/how-are-word-transcript-level-confidence-scores-calculated.mdx new file mode 100644 index 00000000..e500037c --- /dev/null +++ b/fern/pages/faq/async/how-are-word-transcript-level-confidence-scores-calculated.mdx @@ -0,0 +1,7 @@ +--- +title: "How are word/transcript level confidence scores calculated?" +--- + +The confidence scores are derived from the combination of the [softmax layer](https://en.wikipedia.org/wiki/Softmax_function) from the acoustic model and language model probabilities. The models produce probabilities for each predicted word, and we use those probabilities as confidence scores. The transcript level confidence score is calculated by taking the average of all word-level confidence scores. + +Confidence scores fall between `0.0` (low confidence) and `1.0` (high confidence). diff --git a/fern/pages/faq/async/how-can-i-integrate-assemblyai-with-other-services-.mdx b/fern/pages/faq/async/how-can-i-integrate-assemblyai-with-other-services-.mdx new file mode 100644 index 00000000..add8b1ba --- /dev/null +++ b/fern/pages/faq/async/how-can-i-integrate-assemblyai-with-other-services-.mdx @@ -0,0 +1,7 @@ +--- +title: "How can I integrate AssemblyAI with other services?" +--- + +We have existing integrations with services such as Zapier, LangChain, AWS, and more! You can find our existing integrations with other tools and services in our Documentation [here](/docs/integrations) + +Using AssemblyAI API involves making GET and POST requests so as long as you can make those you can use our API with any language, framework, or library. diff --git a/fern/pages/faq/async/how-can-i-make-certain-words-more-likely-to-be-transcribed.mdx b/fern/pages/faq/async/how-can-i-make-certain-words-more-likely-to-be-transcribed.mdx new file mode 100644 index 00000000..28a8699b --- /dev/null +++ b/fern/pages/faq/async/how-can-i-make-certain-words-more-likely-to-be-transcribed.mdx @@ -0,0 +1,7 @@ +--- +title: "How can I make certain words more likely to be transcribed?" +--- + +You can include words, phrases, or both in the `word_boost` parameter. Any term included has its likelihood of being transcribed boosted. + +For more information, check out our docs on [Custom Vocabulary](/docs/speech-to-text/pre-recorded-audio/custom-vocabulary). diff --git a/fern/pages/faq/async/how-can-i-test-assemblyai-without-writing-code-.mdx b/fern/pages/faq/async/how-can-i-test-assemblyai-without-writing-code-.mdx new file mode 100644 index 00000000..390eeeb6 --- /dev/null +++ b/fern/pages/faq/async/how-can-i-test-assemblyai-without-writing-code-.mdx @@ -0,0 +1,11 @@ +--- +title: "How can I test AssemblyAI without writing code?" +--- + +Start exploring our AI capabilities instantly using the [AssemblyAI Playground](https://www.assemblyai.com/playground). This tool lets you test transcription, summarization, topic detection, and content moderation by simply uploading an audio file to a web interface. Note that the Playground is a public demo tool with limited functionality, so avoid uploading sensitive data. + +## Ready to build your app? + +Create powerful AI applications without coding using AssemblyAI's integrations with popular no-code platforms like Zapier, [Make.com](http://Make.com), Power Automate, and Bubble. These platforms make it easy to build automated workflows and applications using our speech AI capabilities. + +[See all of our no-code integrations on our docs](/docs/integrations/no-code) diff --git a/fern/pages/faq/async/how-can-i-transcribe-youtube-videos.mdx b/fern/pages/faq/async/how-can-i-transcribe-youtube-videos.mdx new file mode 100644 index 00000000..f903380b --- /dev/null +++ b/fern/pages/faq/async/how-can-i-transcribe-youtube-videos.mdx @@ -0,0 +1,10 @@ +--- +title: "How can I transcribe YouTube videos?" +--- + +YouTube URLs are not supported in the `audio_url` parameter since it requires a direct link to a downloadable audio file. To transcribe a YouTube video: + +1. Download the video's audio +2. Upload the audio file to our API + +See [How to transcribe YouTube videos cookbook](/docs/guides/transcribe_youtube_videos) (Python) for more information. diff --git a/fern/pages/faq/async/how-can-i-use-universal-1.mdx b/fern/pages/faq/async/how-can-i-use-universal-1.mdx new file mode 100644 index 00000000..414453e2 --- /dev/null +++ b/fern/pages/faq/async/how-can-i-use-universal-1.mdx @@ -0,0 +1,7 @@ +--- +title: "How can I use Universal-2?" +--- + +At AssemblyAI, we use a combination of models to produce your results. Our Best tier will house our most powerful and accurate models, including Universal-2. This tier is best suited for use cases where accuracy is paramount, and end-users will interact directly with the results generated from our models. + +Our Best tier is used by default if no speech model tier is specified. Universal 2 part of our Best tier. There is no need to do anything special to use this new model. Just submit requests as normal and they will always be routed to our latest models. diff --git a/fern/pages/faq/async/how-do-i-enable-universal-2.mdx b/fern/pages/faq/async/how-do-i-enable-universal-2.mdx new file mode 100644 index 00000000..526071d0 --- /dev/null +++ b/fern/pages/faq/async/how-do-i-enable-universal-2.mdx @@ -0,0 +1,5 @@ +--- +title: "How do I enable Universal-2?" +--- + +Our latest model (Universal 2) is automatically enabled for all English transcription requests on our Best tier. You don't need to do anything special - just submit your requests as normal and you'll always use our most advanced speech-to-text model. diff --git a/fern/pages/faq/async/how-do-i-generate-subtitles.mdx b/fern/pages/faq/async/how-do-i-generate-subtitles.mdx new file mode 100644 index 00000000..ed3dba6d --- /dev/null +++ b/fern/pages/faq/async/how-do-i-generate-subtitles.mdx @@ -0,0 +1,9 @@ +--- +title: "How do I generate subtitles?" +--- + +AssemblyAI's API supports exporting transcripts as SRT or VTT files, which are standard subtitle formats used for video captioning. You can learn how to generate subtitles using: + +- [Python Code Example in our Cookbook](/docs/guides/subtitles) +- [Step-by-step Tutorial on our YouTube channel](https://www.youtube.com/watch?v=GgxoRS1qn7w) +- [Code examples in various programming languages on our documentation](/docs/speech-to-text/speech-recognition#export-srt-or-vtt-caption-files) diff --git a/fern/pages/faq/async/how-does-assemblyai-compare-to-other-asr-providers.mdx b/fern/pages/faq/async/how-does-assemblyai-compare-to-other-asr-providers.mdx new file mode 100644 index 00000000..4e1999b4 --- /dev/null +++ b/fern/pages/faq/async/how-does-assemblyai-compare-to-other-asr-providers.mdx @@ -0,0 +1,5 @@ +--- +title: "How does AssemblyAI compare to other ASR providers?" +--- + +Our state-of-the-art speech recognition models are trained on extremely large amounts of data to achieve near human-level performance and robustness across a variety of data. As a result, our models typically perform as good or better than the competition. See [our latest benchmark data](https://www.assemblyai.com/benchmarks) to see how we compare to other ASR providers (including AWS, Google, and more) on a wide range of file types. diff --git a/fern/pages/faq/async/how-does-language-detection-work-for-transcriptions.mdx b/fern/pages/faq/async/how-does-language-detection-work-for-transcriptions.mdx new file mode 100644 index 00000000..a34c9543 --- /dev/null +++ b/fern/pages/faq/async/how-does-language-detection-work-for-transcriptions.mdx @@ -0,0 +1,11 @@ +--- +title: "How does Automatic Language Detection work?" +--- + +Our Automatic Language Detection (ALD) model analyzes samples of the audio to determine the language spoken. It randomly selects up to 3 clips of 30 seconds each from the middle 50% of the audio duration (between 25% and 75% of the total length). + +These 3 clips are passed through our ALD model, which predicts the language probabilities for each clip. The probabilities are then averaged across the clips, and the languages are sorted by their average probability scores. + +This approach helps ensure that the language detection is based on a representative sample of the audio, rather than just the beginning or end portions which may contain greetings, silence, or other non-representative speech. + +If you are seeing low confidence scores for transcriptions in a particular language, it may be due to factors like background noise, accents, or audio quality issues. Our transcription models perform best with clear audio recorded in a quiet environment. diff --git a/fern/pages/faq/async/how-does-the-api-handle-files-that-contain-spoken-audio-in-multiple-languages.mdx b/fern/pages/faq/async/how-does-the-api-handle-files-that-contain-spoken-audio-in-multiple-languages.mdx new file mode 100644 index 00000000..6c8d333d --- /dev/null +++ b/fern/pages/faq/async/how-does-the-api-handle-files-that-contain-spoken-audio-in-multiple-languages.mdx @@ -0,0 +1,9 @@ +--- +title: "How does the API handle files that contain spoken audio in multiple languages?" +--- + +Each of our language models are trained on that specific language. Generally speaking, those models will only transcribe audio spoken in that specific language. There can be some edge cases where the language model is able to transcribe a few words that are in a different language. This can happen when the model may have learned those words due to them being present in the training data. + +For instance, the Spanish model is trained with Spanish audio, but it is not uncommon that some English words are used when speaking Spanish. So when English words appear in the training data used for the Spanish model, the Spanish model can recognize and transcribe those words. + +It's important to note that a language model transcribing words from a different language would be an edge case and not behavior that should be reliably counted on. The expectation is that whatever language model the file is submitted to, will transcribe the spoken audio for that language and will ignore any other spoken languages in the audio. diff --git a/fern/pages/faq/async/how-long-does-it-take-to-transcribe-a-file.mdx b/fern/pages/faq/async/how-long-does-it-take-to-transcribe-a-file.mdx new file mode 100644 index 00000000..5441fe09 --- /dev/null +++ b/fern/pages/faq/async/how-long-does-it-take-to-transcribe-a-file.mdx @@ -0,0 +1,13 @@ +--- +title: "How long does it take to transcribe a file?" +--- + +Processing times for our asynchronous transcription API are based on the duration of the submitted audio and models enabled in the request but the vast majority of files sent to our API will complete in **under 45 seconds**, and with a Real-Time-Factor (RTF) as low as .008x. + +To put an RTF of .008x into perspective, this means you can convert a: + +- 1h3min (75MB) meeting in 35 seconds +- 3h15min (191MB) podcast in 133 seconds +- 8h21min (464MB) video course in 300 seconds + +Files submitted for Streaming Speech-to-Text receive a response within a few hundred milliseconds. diff --git a/fern/pages/faq/async/i-am-getting-an-error-what-should-i-do.mdx b/fern/pages/faq/async/i-am-getting-an-error-what-should-i-do.mdx new file mode 100644 index 00000000..0f6edc67 --- /dev/null +++ b/fern/pages/faq/async/i-am-getting-an-error-what-should-i-do.mdx @@ -0,0 +1,28 @@ +--- +title: "What should I do if I'm getting an error?" +--- + +Any time you make a request to the API, you should receive a JSON response. If you don't receive the expected output, the JSON contains an `error` key with a message value describing the error. + +To retrieve the error from our API: + +1. Make a GET request to fetch the transcript details: + +``` +curl https://api.assemblyai.com/v2/transcript/ \ + -H "Authorization: " +``` + +The response will include an `error` key with details about what went wrong: + +``` +{ + "status": "error", + "error": "Download error, unable to access file at https://example.com/audio.mp3", + ... +} +``` + +2. Check our [Common Errors Guide](/docs/guides/common_errors_and_solutions) for solutions to frequently encountered errors and how to resolve them. + +3. Still stuck? Contact our support team at [support@assemblyai.com](mailto:support@assemblyai.com). Please include your transcript ID, the audio file/URL you're trying to transcribe, and the JSON response from our server so we can help you faster. diff --git a/fern/pages/faq/async/is-there-a-postman-collection-for-using-the-api.mdx b/fern/pages/faq/async/is-there-a-postman-collection-for-using-the-api.mdx new file mode 100644 index 00000000..e5bfd4e8 --- /dev/null +++ b/fern/pages/faq/async/is-there-a-postman-collection-for-using-the-api.mdx @@ -0,0 +1,5 @@ +--- +title: "Is there a Postman collection for using the API?" +--- + +There is! You can find more information on our Postman collection in [this section](/docs/integrations/postman) of our official documentation. See [this blog post](https://www.assemblyai.com/blog/postman-collection) to learn more. diff --git a/fern/pages/faq/async/is-there-a-way-for-us-to-send-the-start-time-end-time-for-transcription-instead-of-transcribing-the-whole-length-of-a-call-recording.mdx b/fern/pages/faq/async/is-there-a-way-for-us-to-send-the-start-time-end-time-for-transcription-instead-of-transcribing-the-whole-length-of-a-call-recording.mdx new file mode 100644 index 00000000..9f78fa93 --- /dev/null +++ b/fern/pages/faq/async/is-there-a-way-for-us-to-send-the-start-time-end-time-for-transcription-instead-of-transcribing-the-whole-length-of-a-call-recording.mdx @@ -0,0 +1,11 @@ +--- +title: "Is there a way for us to send the start time / end time for transcription instead of transcribing the whole length of a call recording?" +--- + +Yes, you can! + +There are parameters that you can add to your `POST` request that allow you to set the start time and end time of where you would like to start/stop transcribing your audio file. + +More specifically you can use the `audio_start_from` parameter and enter the time in milliseconds of where you would like to start transcribing your media file and the `audio_end_at` parameter and enter the time in milliseconds of where you'd like to stop transcribing your media file. + +The parameters are referenced in the [Set the start and end of the transcript](/docs/speech-to-text/pre-recorded-audio/set-the-start-and-end-of-the-transcript) section of our documentation. diff --git a/fern/pages/faq/async/is-there-an-openapi-spec-schema-for-the-api.mdx b/fern/pages/faq/async/is-there-an-openapi-spec-schema-for-the-api.mdx new file mode 100644 index 00000000..21ba77c0 --- /dev/null +++ b/fern/pages/faq/async/is-there-an-openapi-spec-schema-for-the-api.mdx @@ -0,0 +1,5 @@ +--- +title: "Is there an OpenAPI spec/schema for the API?" +--- + +We do have an OpenAPI spec which can be found here: [https://github.com/AssemblyAI/assemblyai-api-spec/blob/main/openapi.yml](https://github.com/AssemblyAI/assemblyai-api-spec/blob/main/openapi.yml) diff --git a/fern/pages/faq/async/read-operation-timed-out-error.mdx b/fern/pages/faq/async/read-operation-timed-out-error.mdx new file mode 100644 index 00000000..52873c19 --- /dev/null +++ b/fern/pages/faq/async/read-operation-timed-out-error.mdx @@ -0,0 +1,14 @@ +--- +title: '"read operation timed out" error' +--- + +This error is a byproduct of a request taking longer than the default timeout value built into the Python SDK, which is 15 seconds. + +This shouldn't be a common occurrence but to allow for the occasional slowdown this setting can be changed like this: + +```python +import assemblyai as aai + +aai.settings.api_key = f"{ASSEMBLYAI_API_KEY}" +aai.settings.http_timeout = 30.0 +``` diff --git a/fern/pages/faq/async/should-i-use-speaker-labels-or-multi-channel.mdx b/fern/pages/faq/async/should-i-use-speaker-labels-or-multi-channel.mdx new file mode 100644 index 00000000..76b38b91 --- /dev/null +++ b/fern/pages/faq/async/should-i-use-speaker-labels-or-multi-channel.mdx @@ -0,0 +1,31 @@ +--- +title: "Should I use Speaker Labels or Multi-channel?" +--- + +The choice depends on your audio source: + +### Use Multi-channel when: + +You have audio where each speaker is recorded on a separate channel, such as: + +- Twilio calls with configured channel separation +- Call center software that exports multiple audio channels +- Any recording where speakers are pre-separated into channels + +Multichannel is more accurate since each speaker's audio is processed independently. + + + Multichannel audio will not automatically separate multiple speakers on the + same channel. If you have multiple speakers on a single channel, speaker + labels are required for speaker separation. + + +### Use Speaker Labels when: + +You have single-channel audio with multiple speakers, such as: + +- Standard audio recordings with multiple participants +- Calls recorded as a single track +- Any audio where speakers share the same channel + +For speaker labels, you can improve accuracy by specifying the expected number of speakers in advance [using the `speakers_expected` parameter](/docs/api-reference/transcripts/submit#request.body.speakers_expected), or set a range of possible speakers [using the `speaker_options` parameter](/docs/speech-to-text/pre-recorded-audio/speaker-diarization#set-a-range-of-possible-speakers). diff --git a/fern/pages/faq/async/should-i-use-the-na-or-eu-endpoint.mdx b/fern/pages/faq/async/should-i-use-the-na-or-eu-endpoint.mdx new file mode 100644 index 00000000..ee54bc43 --- /dev/null +++ b/fern/pages/faq/async/should-i-use-the-na-or-eu-endpoint.mdx @@ -0,0 +1,9 @@ +--- +title: "Should I use the NA or EU endpoint for my Speech-to-Text requests?" +--- + +By default, Speech-to-Text requests using pre-recorded audio are sent to our API via the `https://api.assemblyai.com/v2/transcript` endpoint. We also have an option EU endpoint, `https://api.eu.assemblyai.com/v2/transcript` endpoint that can be used for Speech-to-Text requests using pre-recorded audio. + +Users should only use this EU endpoint when they have data residency needs that require that data be stored in the EU. + +Otherwise, users should use the default NA endpoint as it will generally result in the fastest processing time regardless of what region the user is in. diff --git a/fern/pages/faq/async/what-are-the-recommended-options-for-audio-noise-reduction.mdx b/fern/pages/faq/async/what-are-the-recommended-options-for-audio-noise-reduction.mdx new file mode 100644 index 00000000..692dcd06 --- /dev/null +++ b/fern/pages/faq/async/what-are-the-recommended-options-for-audio-noise-reduction.mdx @@ -0,0 +1,22 @@ +--- +title: What are the recommended options for audio noise reduction? +--- + +When working with audio files, you may want to reduce background noise to improve the listening experience for your users. There are different solutions available depending on your specific needs and resources. + +There are two main approaches you can take for audio noise reduction: + +## 1. Commercial Solution + +[Krisp](https://krisp.ai/developers/#technologies) is a recommended commercial solution for background noise reduction. They offer an SDK that can be integrated into your application. + +## 2. Open Source Solution + +If you prefer more customization and control over processing costs, you can use open-source solutions such as the [noisereduce](https://github.com/timsainb/noisereduce) library. + +For implementation examples with real-time audio using the open-source approach, you can refer to our [streaming API cookbook](https://www.assemblyai.com/docs/guides/noise_reduction_streaming). + + + If you're using our transcription service, we recommend applying noise + reduction for playback purposes only, not before transcription. + diff --git a/fern/pages/faq/async/what-audio-and-video-file-types-are-supported-by-your-api.mdx b/fern/pages/faq/async/what-audio-and-video-file-types-are-supported-by-your-api.mdx new file mode 100644 index 00000000..29b85093 --- /dev/null +++ b/fern/pages/faq/async/what-audio-and-video-file-types-are-supported-by-your-api.mdx @@ -0,0 +1,37 @@ +--- +title: What audio and video file types are supported by your API? +--- + +The AssemblyAI API supports most common audio and video file formats. We recommend that you submit your audio in its native format without additional transcoding or file conversion. Transcoding or converting it to another format can sometimes result in a loss of quality, especially if you're converting compressed formats like `.mp3`. The AssemblyAI API converts all files to 16khz uncompressed audio as part of our transcription pipeline. + +Note that when you upload a video to our API, the audio will be extracted from it and processed independently, so the list of supported video formats isn't exhaustive. + +| Supported audio file types | Supported video file types | +| -------------------------- | --------------------------- | +| .3ga | .webm | +| .8svx | .mts, .m2ts, .ts | +| .aac | .mov | +| .ac3 | .mp2 | +| .aif | .mp4, .m4p (with DRM), .m4v | +| .aiff | .mxf | +| .alac | | +| .amr | | +| .ape | | +| .au | | +| .dss | | +| .flac | | +| .flv | | +| .m4a | | +| .m4b | | +| .m4p | | +| .m4r | | +| .mp3 | | +| .mpga | | +| .ogg, .oga, .mogg | | +| .opus | | +| .qcp | | +| .tta | | +| .voc | | +| .wav | | +| .wma | | +| .wv | | diff --git a/fern/pages/faq/async/what-ip-address-should-i-whitelist-for-assemblyai.mdx b/fern/pages/faq/async/what-ip-address-should-i-whitelist-for-assemblyai.mdx new file mode 100644 index 00000000..30065ffc --- /dev/null +++ b/fern/pages/faq/async/what-ip-address-should-i-whitelist-for-assemblyai.mdx @@ -0,0 +1,12 @@ +--- +title: "What IP Address Should I Whitelist for AssemblyAI?" +--- + +AssemblyAI sends all webhook deliveries from fixed IP addresses: + +| Region | IP Address | +| ------ | ------------ | +| US | 44.238.19.20 | +| EU | 54.220.25.36 | + +When retrieving audio files submitted by your account, our systems use dynamic IP addresses by default. If your network security requirements require a static IP address for these retrieval requests, please [contact us](https://assembly.ai/support) to enable this feature for your account. diff --git a/fern/pages/faq/async/what-is-the-difference-between-custom-vocabulary-and-custom-spelling.mdx b/fern/pages/faq/async/what-is-the-difference-between-custom-vocabulary-and-custom-spelling.mdx new file mode 100644 index 00000000..b1262d06 --- /dev/null +++ b/fern/pages/faq/async/what-is-the-difference-between-custom-vocabulary-and-custom-spelling.mdx @@ -0,0 +1,9 @@ +--- +title: "What is the difference between Custom Vocabulary and Custom Spelling?" +--- + +The **Custom Vocabulary** feature allows you to submit a list of words or phrases to boost the likelihood that the model predicts those words. This is intended to help with words or terms that might be under-represented in the training data. + +The **Custom Spelling** feature allows you to control how words are spelled or formatted in the transcript text. It works like a find and replace feature — anytime you would see X in the API output, it'll be replaced with Y. + +To put it another way, Custom Vocab affects how the model predicts words while Custom Spelling affects words once the transcript text has been created. diff --git a/fern/pages/faq/async/what-is-the-difference-between-the-best-and-nano-tiers.mdx b/fern/pages/faq/async/what-is-the-difference-between-the-best-and-nano-tiers.mdx new file mode 100644 index 00000000..15816640 --- /dev/null +++ b/fern/pages/faq/async/what-is-the-difference-between-the-best-and-nano-tiers.mdx @@ -0,0 +1,7 @@ +--- +title: "What is the difference between the Slam-1, Universal, and Nano models?" +--- + +AssemblyAI offers several state-of-the-art speech recognition models, each optimized for different use cases so that you can choose the model that best fits your needs based on accuracy, latency, cost, and language requirements. + +See [our Models page](https://www.assemblyai.com/docs/getting-started/models) to learn more about these different models. diff --git a/fern/pages/faq/async/what-is-the-minimum-audio-duration-that-the-api-can-transcribe.mdx b/fern/pages/faq/async/what-is-the-minimum-audio-duration-that-the-api-can-transcribe.mdx new file mode 100644 index 00000000..a84b1e32 --- /dev/null +++ b/fern/pages/faq/async/what-is-the-minimum-audio-duration-that-the-api-can-transcribe.mdx @@ -0,0 +1,22 @@ +--- +title: "What is the minimum audio duration that the API can transcribe?" +--- + +The minimum audio duration for a file submitted to our API is 160ms. + +Files submitted to our API that are shorter than 160ms will return an error: + +```json +{ + ... + "status": "error", + "audio_url": "https://foo.bar", + "error": "Audio duration is too short.", + ... +} +``` + +**To fix this issue**: + +- **Add error handling for this error message**: When this error occurs, handle it safely by checking the error string and returning the error. +- **Add pre-submit checks for the duration of the audio file**: Prior to submitting a file for transcription, check the duration using a tool like soxi (part of the SoX package): `soxi -D audio.mp3` diff --git a/fern/pages/faq/async/what-is-the-recommended-file-type-for-using-your-api.mdx b/fern/pages/faq/async/what-is-the-recommended-file-type-for-using-your-api.mdx new file mode 100644 index 00000000..e0022284 --- /dev/null +++ b/fern/pages/faq/async/what-is-the-recommended-file-type-for-using-your-api.mdx @@ -0,0 +1,9 @@ +--- +title: "What is the recommended file type for using your API?" +--- + +Our API supports a wide range of file types so we typically recommend customers submit their audio in whatever its raw format is without any type of transcoding or file conversion. The reason for this is that transcoding or converting it to another format can sometimes result in a loss of quality. + +Some formats like `.wav` are lossless while others like `.mp3` and `.acc` are lossy. See [this article](https://higherhz.com/lossless-vs-lossy-compression-audio-formats/) to learn more. This difference between lossless and lossy is part of the reason we suggest people submit their audio as is without any modification. + +We have transcoding on our end that will take the incoming files and convert them to 16khz `.wav` files so regardless of the file type submitted they will be able to work with our transcription pipeline. diff --git a/fern/pages/faq/async/what-is-universal-2.mdx b/fern/pages/faq/async/what-is-universal-2.mdx new file mode 100644 index 00000000..5d74bfac --- /dev/null +++ b/fern/pages/faq/async/what-is-universal-2.mdx @@ -0,0 +1,15 @@ +--- +title: "What is Universal-2?" +--- + +Universal-2 is our most powerful and accurate next-generation Speech AI model. + +What makes Universal-2 special isn't just better accuracy - it's solving the hardest challenges in conversational AI. From handling messy real-world speech to accurately capturing proper nouns, phone numbers, and formatting, Universal-2 delivers transcripts that are both accurate and clean. + +Here are some metrics for this new model: + +- 24% better at handling proper nouns +- 21% improvement in alphanumeric accuracy +- 15% enhanced text formatting +- Maintains 30% reduction in hallucinations versus other speech-to-text models like Whisper +- 73% of users prefer Universal-2 outputs compared to Universal-1 diff --git a/fern/pages/faq/async/what-languages-do-you-support-.mdx b/fern/pages/faq/async/what-languages-do-you-support-.mdx new file mode 100644 index 00000000..cf647090 --- /dev/null +++ b/fern/pages/faq/async/what-languages-do-you-support-.mdx @@ -0,0 +1,53 @@ +--- +title: "What languages do you support?" +--- + +This page shows which languages are supported by the AssemblyAI Async API, their `language_code` values, and the features available for that language. + +AssemblyAI offers two different speech-to-text models: **Slam-1** and **Universal**. + +Below, you'll find the supported languages for each model, as well as instructions on how to change the language for your transcription. + +## Supported languages for Slam-1 + +