Replies: 4 comments 8 replies
-
I'm currently implementing this idea on aosp. Where do you want the text to go after it is transcribed? What is the user experience you are trying to build? |
Beta Was this translation helpful? Give feedback.
-
So I decided to not use the stream for my app, instead I am recording audio of 10seconds and putting them in a queue where I am transcribing them one at a time. So any idea why when the app is running it is not transcribing the audio, and wait until the app goes to the background before? Also to make sure of the behavior, I installed tour demo app on the same device and it is thz same behavior when I click on the TranscribSample button, it reads the audio, and then it doesn't show anything, when I click the back button and the app goes to the background and I open the app again, then I see the transcribed text on the screen. |
Beta Was this translation helpful? Give feedback.
-
I don't know why, but you can put logs in between your method calls. I am
using android.telecom to make calls
…On Tue, Mar 21, 2023 at 6:40 PM usfaa444 ***@***.***> wrote:
So I decided to not use the stream for my app, instead I am recording
audio of 10seconds and putting them in a queue where I am transcribing them
one at a time.
The problem I encountered is the following.
When I test on a mibile phone, like my Samsung galaxy s22 ultra,
everything is fine, and i have every audio transcribed with 3 to 4 seconds.
But on some other devices, when transcribing the first audio, the
transcsibData funtion in the LibWhisper get stuck it doesn't return
anything until I close the app (the app goes to the background) and then at
this momebt, then transcrib result is returned and the other audio in the
queue get transcribed as well with 3 to 4 seconds.
So any idea why when the app is running it is not transcribing the audio,
and wait until the app goes to the background before?
Also to make sure of the behavior, I installed tour demo app on the same
device and it is thz same behavior when I click on the TranscribSample
button, it reads the audio, and then it doesn't show anything, when I click
the back button and the app goes to the background and I open the app
again, then I see the transcribed text on the screen.
—
Reply to this email directly, view it on GitHub
<#560 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/A446LP6AUOHQHX5QSP5ICZLW5JKA3ANCNFSM6AAAAAAVOX5A34>
.
You are receiving this because you commented.Message ID:
***@***.***>
|
Beta Was this translation helpful? Give feedback.
-
I use whisper.cpp from my de-googled Android phone daily for months now for dictation to SMS and email and web search. My preferred method is to use the Konele app as the voice input frontend which sends the audio to my own private server which is running a very simple python wrapper around whisper.cpp. Here is my project: https://github.com/rpdrewes/whisper-websocket-server Another alternative is this port of whisper.cpp to run natively on Android. I have tested it and it also works well but it is not as fast as my server-based approach in my environment (but if you have a slower server and/or a faster phone it might be better for you). It uses Konele as a frontend for the input as well, I believe. See: https://github.com/alex-vt/WhisperInput |
Beta Was this translation helpful? Give feedback.
-
So I was going through this awesome library, and by the way thank you very much for such a great work, I am sure it will help thousands of developers, including me.
So I want to use the streaming to be able to transcribe any speech detected from the microphone, I saw that the stream demo you gave in the video, you typed a command in the CLI, is it possible to have this functionality implemented in Android please?
Beta Was this translation helpful? Give feedback.
All reactions