-
Notifications
You must be signed in to change notification settings - Fork 2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Speech client v2, my, working OK, example #3578
Comments
Does this still work for you? I created my recognizer in cloud console and am just specifying it as a string. Im however getting the error "Error: 3 INVALID_ARGUMENT: Malordered Data Received. Expected audio but none was set. Send exactly one config, followed by audio data.". I really cant figure out what im doing wrong... Edit: @sorokinvj I fixed it so the audio data can send but get back no response from STT. The stream ends up timing out. I have interimResults to true. |
Yeah, all of this is because you are somehow sending wrong audio data, if you share your code I might get more insight. @adambeer sorry, I saw it just now. My code runs production server, no problem so far |
I get code: 7, when I execute recognizeStream.write(streamingRecognizeRequest); This my code:
const streamingRecognitionConfig: google.cloud.speech.v2.IStreamingRecognitionConfig = const streamingRecognizeRequest: google.cloud.speech.v2.IStreamingRecognizeRequest = const recognizeStream = client While a non-streaming request works well: const request: RecognizeBatchRequest = { Any idea why? the service account i use has cloud speech admin permissions. PS: sorry for the non-code formatting. |
I am getting Error: 3 INVALID_ARGUMENT: Invalid resource field value in the request. import { v2 as speech } from '@google-cloud/speech'; const speechClient = new speech.SpeechClient(); const request = { const recognizeStream = speechClient.streamingRecognize(request) NOTE: I am on node 18. |
I don't know, man, sorry. Your code is doing something else than mine, but to understand your usecase I would need to see more than what you've shared.
|
Can you try to not use any model and leave this field undefined?
Ensure FIREBASE_PROJECT_ID is set
It might be useful to have a basic check in the code, that you send this request only once in the first packet, though I doubt it has anything to do with permission issue. Sorry, have no idea why you might stuck with this. |
Thank you for your time. I got the config set correctly. I was also able to get a stream from a local file to transcribe correctly. However, I am getting no response on this.
This is a base64 encoded payload coming from a websocket message. The config sends properly but there is clear some sort of problem with the buffer and audio data. I verified the incoming data and that the data is sent to _streamingRecognize. These lines worked in v1 by the way. No error codes. Just no response then timeout after cancelling the project. UPDATE: I just realized the other is only giving a response on stream end. I think this is my problem. this needs to transcribe voice data via a websocket. |
@sorokinvj Im calling it from a google function onCall with the code below and Im receiving I tried without the model, but still the same problem
|
UPDATE
|
I can get transcriptions but the problem I am having is that transcriptions only come back after the stream has ended |
I know the models offer mulaw encoding but everytime i set mulaw with a recognizer in console it flips back to linear16. Anyone able to help? |
Hey guys, I am not sure where to put it, but I just want to share my implementation of the speech client v2 and some thoughts about migrating from v1 to v2.
Official docs are unfortunately provide only python code, and there is not much info except this repo and the example I used to migrate: transcribeStreaming.v2.js
My sdk verison is:
In my code I first initialize the service as:
and then use
service.transcribeAudio(data)
whenever there is a new audio coming from the frontend, which usesthus an audio chunk is just a browser
Blob
object.My service:
Migration considerations
recognizingClient.write(audioData)
before, now you should do (but only once!)recognizingClient.write(newConfigWithRecognizer)
and thenrecognizingClient.write({ audio: audioData })
<<< notice the object notation_streamingRecognize()
(this probably is likely to be changed)The text was updated successfully, but these errors were encountered: