From 8e7d6cba4fbfca255b0ae49dc17bf673a130948c Mon Sep 17 00:00:00 2001 From: Alexandre Lissy Date: Mon, 10 Aug 2020 20:32:19 +0200 Subject: [PATCH] Bump VERSION to 0.8.1 --- doc/USING.rst | 8 ++++---- doc/index.rst | 12 ++++++------ native_client/dotnet/DeepSpeechWPF/App.xaml.cs | 2 +- .../deepspeech_ios_test/SpeechRecognitionImpl.swift | 4 ++-- training/deepspeech_training/VERSION | 2 +- 5 files changed, 14 insertions(+), 14 deletions(-) diff --git a/doc/USING.rst b/doc/USING.rst index 316041e1a3..779c4588d7 100644 --- a/doc/USING.rst +++ b/doc/USING.rst @@ -37,8 +37,8 @@ If you want to use the pre-trained English model for performing speech-to-text, .. code-block:: bash - wget https://github.com/mozilla/DeepSpeech/releases/download/v0.8.0/deepspeech-0.8.0-models.pbmm - wget https://github.com/mozilla/DeepSpeech/releases/download/v0.8.0/deepspeech-0.8.0-models.scorer + wget https://github.com/mozilla/DeepSpeech/releases/download/v0.8.1/deepspeech-0.8.1-models.pbmm + wget https://github.com/mozilla/DeepSpeech/releases/download/v0.8.1/deepspeech-0.8.1-models.scorer There are several pre-trained model files available in official releases. Files ending in ``.pbmm`` are compatible with clients and language bindings built against the standard TensorFlow runtime. Usually these packages are simply called ``deepspeech``. These files are also compatible with CUDA enabled clients and language bindings. These packages are usually called ``deepspeech-gpu``. Files ending in ``.tflite`` are compatible with clients and language bindings built against the `TensorFlow Lite runtime `_. These models are optimized for size and performance in low power devices. On desktop platforms, the compatible packages are called ``deepspeech-tflite``. On Android and Raspberry Pi, we only publish TensorFlow Lite enabled packages, and they are simply called ``deepspeech``. You can see a full list of supported platforms and which TensorFlow runtime is supported at :ref:`supported-platforms-inference`. @@ -136,7 +136,7 @@ Note: the following command assumes you `downloaded the pre-trained model <#gett .. code-block:: bash - deepspeech --model deepspeech-0.8.0-models.pbmm --scorer deepspeech-0.8.0-models.scorer --audio my_audio_file.wav + deepspeech --model deepspeech-0.8.1-models.pbmm --scorer deepspeech-0.8.1-models.scorer --audio my_audio_file.wav The ``--scorer`` argument is optional, and represents an external language model to be used when transcribing the audio. @@ -200,7 +200,7 @@ Note: the following command assumes you `downloaded the pre-trained model <#gett .. code-block:: bash - ./deepspeech --model deepspeech-0.8.0-models.pbmm --scorer deepspeech-0.8.0-models.scorer --audio audio_input.wav + ./deepspeech --model deepspeech-0.8.1-models.pbmm --scorer deepspeech-0.8.1-models.scorer --audio audio_input.wav See the help output with ``./deepspeech -h`` for more details. diff --git a/doc/index.rst b/doc/index.rst index 7cc572aa51..2dad0e0094 100644 --- a/doc/index.rst +++ b/doc/index.rst @@ -20,15 +20,15 @@ To install and use DeepSpeech all you have to do is: pip3 install deepspeech # Download pre-trained English model files - curl -LO https://github.com/mozilla/DeepSpeech/releases/download/v0.8.0/deepspeech-0.8.0-models.pbmm - curl -LO https://github.com/mozilla/DeepSpeech/releases/download/v0.8.0/deepspeech-0.8.0-models.scorer + curl -LO https://github.com/mozilla/DeepSpeech/releases/download/v0.8.1/deepspeech-0.8.1-models.pbmm + curl -LO https://github.com/mozilla/DeepSpeech/releases/download/v0.8.1/deepspeech-0.8.1-models.scorer # Download example audio files - curl -LO https://github.com/mozilla/DeepSpeech/releases/download/v0.8.0/audio-0.8.0.tar.gz - tar xvf audio-0.8.0.tar.gz + curl -LO https://github.com/mozilla/DeepSpeech/releases/download/v0.8.1/audio-0.8.1.tar.gz + tar xvf audio-0.8.1.tar.gz # Transcribe an audio file - deepspeech --model deepspeech-0.8.0-models.pbmm --scorer deepspeech-0.8.0-models.scorer --audio audio/2830-3980-0043.wav + deepspeech --model deepspeech-0.8.1-models.pbmm --scorer deepspeech-0.8.1-models.scorer --audio audio/2830-3980-0043.wav A pre-trained English model is available for use and can be downloaded following the instructions in :ref:`the usage docs `. For the latest release, including pre-trained models and checkpoints, `see the GitHub releases page `_. @@ -44,7 +44,7 @@ Quicker inference can be performed using a supported NVIDIA GPU on Linux. See th pip3 install deepspeech-gpu # Transcribe an audio file. - deepspeech --model deepspeech-0.8.0-models.pbmm --scorer deepspeech-0.8.0-models.scorer --audio audio/2830-3980-0043.wav + deepspeech --model deepspeech-0.8.1-models.pbmm --scorer deepspeech-0.8.1-models.scorer --audio audio/2830-3980-0043.wav Please ensure you have the required :ref:`CUDA dependencies `. diff --git a/native_client/dotnet/DeepSpeechWPF/App.xaml.cs b/native_client/dotnet/DeepSpeechWPF/App.xaml.cs index d4b87d6e60..e1c1438aef 100644 --- a/native_client/dotnet/DeepSpeechWPF/App.xaml.cs +++ b/native_client/dotnet/DeepSpeechWPF/App.xaml.cs @@ -20,7 +20,7 @@ protected override void OnStartup(StartupEventArgs e) { //Register instance of DeepSpeech DeepSpeechClient.DeepSpeech deepSpeechClient = - new DeepSpeechClient.DeepSpeech("deepspeech-0.8.0-models.pbmm"); + new DeepSpeechClient.DeepSpeech("deepspeech-0.8.1-models.pbmm"); SimpleIoc.Default.Register(() => deepSpeechClient); SimpleIoc.Default.Register(); diff --git a/native_client/swift/deepspeech_ios_test/SpeechRecognitionImpl.swift b/native_client/swift/deepspeech_ios_test/SpeechRecognitionImpl.swift index 89ae411464..f7f9cb15ab 100644 --- a/native_client/swift/deepspeech_ios_test/SpeechRecognitionImpl.swift +++ b/native_client/swift/deepspeech_ios_test/SpeechRecognitionImpl.swift @@ -26,8 +26,8 @@ class SpeechRecognitionImpl : NSObject, AVCaptureAudioDataOutputSampleBufferDele private var audioData = Data() override init() { - let modelPath = Bundle.main.path(forResource: "deepspeech-0.8.0-models", ofType: "tflite")! - let scorerPath = Bundle.main.path(forResource: "deepspeech-0.8.0-models", ofType: "scorer")! + let modelPath = Bundle.main.path(forResource: "deepspeech-0.8.1-models", ofType: "tflite")! + let scorerPath = Bundle.main.path(forResource: "deepspeech-0.8.1-models", ofType: "scorer")! model = try! DeepSpeechModel(modelPath: modelPath) try! model.enableExternalScorer(scorerPath: scorerPath) diff --git a/training/deepspeech_training/VERSION b/training/deepspeech_training/VERSION index a3df0a6959..6f4eebdf6f 100644 --- a/training/deepspeech_training/VERSION +++ b/training/deepspeech_training/VERSION @@ -1 +1 @@ -0.8.0 +0.8.1