azure_stt_flutter 0.0.5 copy "azure_stt_flutter: ^0.0.5" to clipboard
azure_stt_flutter: ^0.0.5 copied to clipboard

A Flutter package for real-time Speech-to-Text transcription using Microsoft Azure Cognitive Services with BLoC/Cubit pattern. Supports Mobile, Desktop, and Web platforms

Azure STT Flutter #

A Flutter package for real-time Speech-to-Text (transcription) using Microsoft Azure Cognitive Services. This library provides a reactive, stream-based API built to easily integrate speech recognition into your Flutter applications.

Features #

  • Real-time Transcription: Receive intermediate results (hypothesis) and finalized text as the user speaks.
  • Cross-Platform: Supports Mobile (iOS, Android), Desktop (macOS, Windows, Linux), and Web.
  • Auto-Silence Timeout: Automatically clears the text after a configurable period of silence.
  • Multi-Language & LID: Supports single-language recognition and multi-language identification (LID).

Language Identification (LID) Modes #

The library supports three main ways to handle spoken languages. Choosing the right mode is critical for performance and accuracy.

1. Single Language (Fastest) #

This is the recommended mode for fastest subtitles and real-time feedback. The engine doesn't spend time identifying the language; it starts transcribing immediately using the provided locale.

How to use: Provide only one language in the list.

final azureStt = AzureSpeechToText(
  subscriptionKey: '...',
  region: '...',
  languages: ['en-US'], // Single language
);

2. At-Start Detection #

The service identifies the language(s) talked at the beginning of the audio and then transcribes using that language for the rest of the session. It supports up to 4 candidate languages.

Note: The first few seconds of audio are used for identification, which might introduce a slight initial delay in transcription.

How to use: Provide up to 4 languages and set languageIdMode to 'AtStart' (default).

final azureStt = AzureSpeechToText(
  subscriptionKey: '...',
  region: '...',
  languages: ['en-US', 'it-IT', 'es-ES', 'fr-FR'],
  languageIdMode: .atStart, // Default
);

3. Continuous Detection #

The service continuously monitors the audio and can switch the transcription language mid-stream if the speaker changes. It supports up to 10 candidate languages.

Note: This mode is the most flexible but requires the service to constantly evaluate the language, which is best for multi-lingual conversations.

How to use: Provide up to 10 languages and set languageIdMode to 'Continuous'.

final azureStt = AzureSpeechToText(
  subscriptionKey: '...',
  region: '...',
  languages: ['en-US', 'it-IT', 'es-ES', 'de-DE', 'pt-PT', 'nb-NO', 'sv-SE', 'uk-UA'],
  languageIdMode: .continuous,
);

Getting Started #

1. Permissions #

Android

Add the microphone permission to android/app/src/main/AndroidManifest.xml:

<uses-permission android:name="android.permission.RECORD_AUDIO" />
<uses-permission android:name="android.permission.INTERNET" />

iOS

Add the microphone usage description to ios/Runner/Info.plist:

<key>NSMicrophoneUsageDescription</key>
<string>This app needs access to the microphone for speech recognition.</string>

macOS

Add the microphone entitlement to macos/Runner/DebugProfile.entitlements and Release.entitlements:

<key>com.apple.security.device.audio-input</key>
<true/>

Usage #

Initialization #

Initialize the AzureSpeechToText instance.

final azureStt = AzureSpeechToText(
  subscriptionKey: 'YOUR_AZURE_KEY',
  region: 'westeurope',
  languages: ['en-US'],
  textClearTimeout: const Duration(seconds: 2),
);

Listening to Updates #

The library exposes a transcriptionStateStream which emits TranscriptionState updates. When using LID, the detectedLanguage field will contain the identified locale.

StreamBuilder<TranscriptionState>(
  stream: azureStt.transcriptionStateStream,
  builder: (context, snapshot) {
    final state = snapshot.data;
    if (state == null) return SizedBox();

    return Column(
      children: [
        if (state.detectedLanguage != null)
          Text('Language: ${state.detectedLanguage}'),
        // Combined text (finalized + intermediate)
        Text(state.text),

        // Or access them separately
        // Text(state.intermediateText), // Changing hypothesis
        // Text(state.finalizedText.join(' ')), // Confirmed sentences
      ],
    );
  },
)

Controls #

// Start listening
await azureStt.startListening()

// Stop listening
azureStt.stopListening()

// Check if listening
azureStt.isListening()

// Dispose when done
azureStt.dispose()

Architecture #

The library is built using the BLoC/Cubit pattern to manage the state of the transcription.

TranscriptionCubit #

The central state manager. It processes events from the Azure Service and emits TranscriptionState.

TranscriptionState #

An immutable object containing:

  • intermediateText: The real-time, changing text (hypothesis) that Azure sends while you are speaking.
  • finalizedText: A list of completed sentences (phrases) that Azure has confirmed.
  • text: A helper field that combines finalized and intermediate text for easier display.
  • detectedLanguage: The BCP-47 locale detected by the service (when using LID).
  • isListening: A boolean indicating if the microphone is active.

Authentication #

The library handles authentication differently depending on the platform due to browser limitations.

Mobile & Desktop #

  • Mechanism: The library uses the Subscription Key to get a short-lived Access Token from Azure.
  • Connection: It connects to the Azure WebSocket URL, passing this token in the HTTP Authorization Header (Authorization: Bearer <token>). This is the standard, secure way.

Web #

  • Limitation: Standard browser WebSocket APIs do not allow setting custom HTTP headers during the handshake.
  • Solution: The library connects to the Azure WebSocket URL but passes the authentication credentials directly in the URL Query Parameters.
  • Security Note: Because query parameters can potentially be logged, using the Token Fetcher approach (generating tokens on your backend) is highly recommended for Web deployments to avoid exposing your long-lived Subscription Key.

License #

This project is licensed under the MIT License - see the LICENSE file for details.

1
likes
150
points
203
downloads

Publisher

unverified uploader

Weekly Downloads

A Flutter package for real-time Speech-to-Text transcription using Microsoft Azure Cognitive Services with BLoC/Cubit pattern. Supports Mobile, Desktop, and Web platforms

Repository (GitHub)
View/report issues

Documentation

API reference

License

MIT (license)

Dependencies

equatable, flutter, flutter_bloc, http, record, uuid, web_socket_channel

More

Packages that depend on azure_stt_flutter