speech-to-text

com/sun/jna/android-arm/libjnidispatch.so not found in resource path

僤鯓⒐⒋嵵緔 提交于 2019-12-23 18:04:30
问题 All of the following is being done in Android Studio. I have successfully compiled and tested the Android Watson Speech to Text demo app. I then created a library project containing the Watson related API's and a 2nd app project with a simple UI that references the Watson library project. The UI successfully starts and calls Watson speech to text api's. I thought I was set to use the Watson library project for real. So I incorporated the Watson API project into my 'real' project. When I start

SpeechRecognizer: no selected voice recognition service

一笑奈何 提交于 2019-12-23 12:53:58
问题 This is how i start my RecogniseListener intent: Intent intent = new Intent(RecognizerIntent.ACTION_RECOGNIZE_SPEECH); intent.putExtra(RecognizerIntent.EXTRA_LANGUAGE_MODEL,RecognizerIntent.LANGUAGE_MODEL_FREE_FORM); intent.putExtra("android.speech.extra.DICTATION_MODE", true); intent.putExtra(RecognizerIntent.EXTRA_CALLING_PACKAGE,this.getPackageName()); intent.putExtra(RecognizerIntent.EXTRA_PARTIAL_RESULTS, true); intent.putExtra(RecognizerIntent.EXTRA_MAX_RESULTS,1); sr.startListening

Python Speech Compare

人盡茶涼 提交于 2019-12-23 09:57:02
问题 I have two .wav files that I need to compare and decide if they contain the same words (same order too). I have been searching for the best method for a while now. I can't figure out how to have pyspeech use a file as input. I've tried getting the CMU sphinx project working but I cant seem to get GStreamer to work with Python 27 let alone their project. I've messed around with DragonFly as well with no luck. I am using Win7 64bit with Python27. Does anyone have any ideas? Any help is greatly

INVALID_ARGUMENT: Request payload size exceeds the limit: 10485760 bytes

时光总嘲笑我的痴心妄想 提交于 2019-12-23 08:08:37
问题 I'm using for the first time the GCS Speech API for a project to convert a series of audio files to text. Each file has around 60 minutes and is a person talking continuously during the whole time. I've installed the GC SDK and I'm using it to perform the requests as shown bellow: gcloud ml speech recognize-long-running \ "/path/to/file/audio.flac" \ --language-code="pt-PT" --async Every time I run this on one of my recording, it gives the following error message: ERROR: (gcloud.ml.speech

INVALID_ARGUMENT: Request payload size exceeds the limit: 10485760 bytes

爱⌒轻易说出口 提交于 2019-12-23 08:08:21
问题 I'm using for the first time the GCS Speech API for a project to convert a series of audio files to text. Each file has around 60 minutes and is a person talking continuously during the whole time. I've installed the GC SDK and I'm using it to perform the requests as shown bellow: gcloud ml speech recognize-long-running \ "/path/to/file/audio.flac" \ --language-code="pt-PT" --async Every time I run this on one of my recording, it gives the following error message: ERROR: (gcloud.ml.speech

Recognition by comparison with pre-recorded words

≡放荡痞女 提交于 2019-12-23 05:51:39
问题 Requirement: I want a system/algorithm that is similar to the call system in the car that you say a name that you want to call, it will find out that name match your voice. My requirement is that I will use pre-recorded voices instead of Name. e.g.: User record target voices for several words in non-English language, e.g.: "Keyboard-JP-01.wav" => Keyboard "Monitor-JP-01.wav" => Monitor After some training or even no training, when the user speak a word again, the speech recognizer should able

Recognition by comparison with pre-recorded words

僤鯓⒐⒋嵵緔 提交于 2019-12-23 05:51:33
问题 Requirement: I want a system/algorithm that is similar to the call system in the car that you say a name that you want to call, it will find out that name match your voice. My requirement is that I will use pre-recorded voices instead of Name. e.g.: User record target voices for several words in non-English language, e.g.: "Keyboard-JP-01.wav" => Keyboard "Monitor-JP-01.wav" => Monitor After some training or even no training, when the user speak a word again, the speech recognizer should able

Android: Arabic speech recognition - offline

百般思念 提交于 2019-12-23 05:34:08
问题 Can any one help me or give me a clue about Arabic speech recognition - offline for android?? I want to develop simple app that do speech recognition - offline as doing this online I know how... if any one have suggestion or link to an open source project or even library to buy that do this efficiently it will be amazing. thanks in advance. 回答1: use intent.putExtra(RecognizerIntent.EXTRA_LANGUAGE,"ar-JO"); The below code worked fine with me: package tk.oryx.voice; import android.app.Activity;

C# system.speech.recognition alternates

ぐ巨炮叔叔 提交于 2019-12-23 05:33:13
问题 I am using using system.speech.recognition to recognize a single word from many. Because the accuracy is not very good, I want to consider more options the engine gives me. However, the e.Result.Alternates only contains one result which is the same as e.Result.word. Is there a ways to force the engine to give me more options? Thank you! 来源: https://stackoverflow.com/questions/8214904/c-sharp-system-speech-recognition-alternates

how to pass language in speech recognition on android apps?

旧时模样 提交于 2019-12-22 13:59:34
问题 I've been working on speech Recognition API in android and found out that the speech results vary allot when the language settings are changed , is there a way to set it programmatically ? or is there an intent to lunch the speech language settings screen ? or what else ? note: I tried to use this intent extra: intent.putExtra(RecognizerIntent.EXTRA_LANGUAGE_PREFERENCE, "en-US"); and Intent detailsIntent = new Intent(RecognizerIntent.ACTION_GET_LANGUAGE_DETAILS); sendOrderedBroadcast