allright, i am making a application as a digital life assistant and i am thinking of moving from the awful windows speech recognition over to Google chrome's new speech engine found here
[login to view URL]
i have done some research and found it is possible to implement this into .NET, however i havn't been able to do this, what i want is for the user to be able to speak into their microphone and the application detects it and when the user stops speaking it sends what ever they said to google's speech api and then returns just the text result to a textbox inside the program with the highest "confidence" rating.
so far i have been able to achieve the application just as a test on my end to send a .flac file of audio google analyses it and returns:
{"status":0,"id":"10ba02de37960b1fb81ce85d9bd69b7f-1","hypotheses":[{"utterance":"hello how are you today","confidence":0.9641105},{"utterance":"hello how are you to day"}]}
so basically, like i said above, i want someone to figure out how to code this so that the person simply talks to the app to initiate the dictation to the app, then the app records what they say, and when it detects that the user has stopped speaking and then sends it to the online api and then returns what the highest probability spoken into a textbox within the program.
i found this site that may be some help
[login to view URL]
I've worked on tons of SOAP / Rest / Json RPC / XML RPC APIs . that includes a lot of APIs of google products, ebay , paypal , freshbooks, quickboos, shopify , magento and a lot of others. You can see my profile to find API projects and their reviews / ratings. This seems an interesting project and I am willing to provide my services. Please check PM for my work samples.
Hello, this should be done by audio capturing module, that can detect whenever user start/stop talking, then record to audio file and send to google speech API.