google-cloud-speech

25s Latency in Google Speech to Text

我只是一个虾纸丫 提交于 2019-12-22 13:53:10
问题 This is a problem I ran into using the Google Speech to Text Engine. I am currently streaming 16 bit / 16 kHz audio real time in 32kB chunks. But there is an average 25 second latency between sending audio and receiving transcripts, defeating the purpose of real time transcription. Why is there such high latency? 回答1: The Google Speech to Text documentation recommends using a 100 ms frame size to minimize latency. 32kB * (8 bits / 1 byte) * ( 1 sample / 16 bits ) * (1 sec / 16000 samples ) =

Difference between Android Speech to Text API (Recognizer Intent) and Google Cloud Speech API?

假如想象 提交于 2019-12-17 22:22:35
问题 So i'm looking into building a speech to text app for fun. I did some research and found an inbuilt Speech to Text API using RecognizerIntent that is free, but also found that google is now offerieng a cloud speech API that the charge for. My question is, what is the difference between them, and If i use the inbuilt RecognizerIntent, is it free? 回答1: For the Google Cloud Speech API, refer to the following link: https://cloud.google.com/speech/. Here are the highlights: It supports 80

Google Speech Recognition API Result is Empty

回眸只為那壹抹淺笑 提交于 2019-12-17 02:49:21
问题 I'm performing an asynchronous request to Google Cloud Speech API, and I do not know how to get the result of operation: Request POST: https://speech.googleapis.com/v1beta1/speech:asyncrecognize Body: { "config":{ "languageCode" : "pt-BR", "encoding" : "LINEAR16", "sampleRate" : 16000 }, "audio":{ "uri":"gs://bucket/audio.flac" } } Which returns: { "name": "469432517" } So, I do a POST: https://speech.googleapis.com/v1beta1/operations/469432517 Which returns: { "name": "469432517", "metadata"

Improving accuracy of Google Cloud Speech API

天涯浪子 提交于 2019-12-13 14:12:32
问题 I am currently recording audio from a web page on my Mac OS computer and running it through the cloud speech api to produce a transcript. However, the results aren't that accurate and there are chunks of missing words in the results. Are there any steps that would help me yield more accurate results? Here are the steps I am taking to convert audio to text: Use Soundflower to channel audio output from my soundcard to mic in. Play audio from website Use quickTime player to record audio which is

How to record microphone audio in JavaScript and submit to DialogFlow?

南笙酒味 提交于 2019-12-13 08:08:50
问题 How can I record audio from the microphone in JavaScript and submit it to DialogFlow, without going through a server? 回答1: There are two parts to this question: How to record microphone audio in a format DialogFlow will understand. How to actually submit that audio to DialogFlow, with proper authentication. Part 1 For recording microphone audio in a format DialogFlow will understand, I use opus-recorder, then convert the blob it returns using the code below: function BlobToDataURL(blob: Blob)

Google Cloud Speech API - certificate verify failed in Python

冷暖自知 提交于 2019-12-13 06:20:35
问题 I'm using SpeechRecognition library. import speech_recognition as sr AUDIO_FILE = 'test_audio.wav' with open("api-key.json") as f: GOOGLE_CLOUD_SPEECH_CREDENTIALS = f.read() r = sr.Recognizer() with sr.AudioFile(AUDIO_FILE) as source: audio = r.record(source) print('Starting recognition...') print(r.recognize_google_cloud(audio, credentials_json=GOOGLE_CLOUD_SPEECH_CREDENTIALS)) print('Completed') When above code is run, an error occurs - ssl.SSLError: [SSL: CERTIFICATE_VERIFY_FAILED]

m4a audio file convert to flac c# [closed]

白昼怎懂夜的黑 提交于 2019-12-13 03:41:26
问题 Closed. This question is off-topic. It is not currently accepting answers. Want to improve this question? Update the question so it's on-topic for Stack Overflow. Closed 6 months ago . I have been searching for the C# library or code sample which could help me to do the conversion of m4a audio file to flac , but more importantly with sample rate (of 16000), mono channel and bit resolution of either 16 or 24. Right now, I have found library solutions like NAudio, MediaToolKit and SOXSharp .

Google cloud speech syncrecognize “INVALID_ARGUMENT”

牧云@^-^@ 提交于 2019-12-12 11:26:43
问题 I have managed the "overview tutorial" : https://cloud.google.com/speech/docs/getting-started Then I tried to use my own audio file . I uploaded a .flac file with a sample rate of 16000Hz. I only changed the sync-request.json file below with my own audio file hosted on google cloud storage ( gs://my-bucket/test4.flac ) { "config": { "encoding":"flac", "sample_rate": 16000 }, "audio": { "uri":"gs://my-bucket/test4.flac" } } The file is well recognized but the request return an "INVALID

Error posting to Clearcut: (null), with Status Code: 400 in Google Cloud Speech API

匆匆过客 提交于 2019-12-12 04:13:53
问题 I want to use Google Cloud Speech API in one of my app. I want to use Cloud Speech Streaming gRPC. I am using following sample as an example. Cloud Speech Streaming gRPC Objective-C Sample I installed this demo application and it is working fine but when I installed required Google Speech libs in my application and tried to run my application, my application crashed with above mentioned error. 回答1: I removed Firebase/Core from pod file and also removed it's headers in the files where I am

Specify Region for Google Speech API?

拜拜、爱过 提交于 2019-12-11 14:47:29
问题 We are using Google Speech API as part of our service. Due to new GDPR rules we have to make sure none of our data leaves the EU. All other services seems to be able to specify a region including Google Cloud Storage. However, I haven't been able to find any documentation related to Google Speech API. Anybody know if it is possible to specify a region for Google Speech API to avoid sending our data outside the EU? 回答1: Found my answer: https://cloud.google.com/about/locations/?region=europe