I'm interested in using Actions and the Assistant to create dynamic dialog for a video game.
Specifically I would want players to be able to speak (literally) to characters and for the characters responses to be determined by Actions, just like the Assistant.
Is there any version of the Assistant available that can be integrated into a game? As far as I can see they offer a lot of the building block services to developers, through the cloud, but nothing as fully featured as Google Assistant
Sounds like a cool scenario. Not something Actions on Google supports directly, but if you want to experiment, you could use the Google Assistant SDK to host the Assistant in your game and respond to queries that are meant for your players.
https://developers.google.com/assistant/sdk/
Love to see what you come up with.
It pretty much comes down to which Framework you use when building your game. If you use Unity for instance, you can use API.AI's Unity SDK.
There are also a lot of other SDKs provided. I don't think you really have to include the complete Google Assistant SDK, since you most likely will want to write your own responses (?). Some SDKs have speech recognition included, for others you will need a Speech Recignition framework, for instance Google Cloud Speech API.
Related
I have problem with real-time voice transcription. In my application people can make a video call each other according to their language choice, I want to show subtitles of conversations in real time to users. I used Agora SDK for video call feature. Is it possible to add real-time transcription?
I found Google Cloud Speech API but there is not enough documentation for dart. I would really be pleased if you give me some starting point
It is just a suggestion. I don’t know if it is valid or not, but you can try. You can apply the Java code or Kotlin or Xamarin, this will help you in setting 'Google Cloud Speech API'.
Excuse me for my explanation, I hope to be of benefit to you
I'm working at my new mobile game. The work is going well but when I tried to use google play services at the game, it didn't work by several reasons that I couldn't understand.
There are many guides about using google play service, but still they are all scattered. So I can't totally understand about this.
Could you help me to how to use google play service step by step?
Or link that can solve my problem.
You may check this link for Google Play Games plugin for Unity. This plugin allows you to access the Google Play Games API through Unity's social interface. The plugin provides support for the following features of the Google Play Games API:
sign in
unlock/reveal/increment achievement
post score to leaderboard
cloud save read/write
show built-in achievement/leaderboards UI
events
video recording of gameplay
nearby connections
turn-based multiplayer
real-time multiplayer
To use the plugin, you must first configure your game in the Google Play Developer Console. Follow the instructions on creating a client ID. Be particularly careful when entering your package name and your certificate fingerprints, since mistakes on those screens can be difficult to recover from.
check it
play-unity-plugins
-play billing
-play instant
-play assets delivery
-anfd so on
I had started out doing a games/apps (just make for fun so doesn't have exciting gameplay) where it can read out the word that show on screen. Had tested out http://translate.google.com/translate_tts?tl=en&q=Hello%20World , but it doesn't work out so I started a trial on Google Cloud Platform, and now I stuck on how to use its Text-To-Speech function on Unity. Is it possible to use Google TTS on Unity? Cause, when I use using Google.Could.Text-To-Speech.V1 it shows error where it doesn't recognize Google.
The text-to-speech API works by making an HTTP POST request to https://texttospeech.googleapis.com/v1/text:synthesize, with the text in the body of the request. I'm not familiar with Unity3d, but a quick check of the documentation shows you can use UnityWebRequest.Post to make a HTTP POST request - note that this means your game must be online.
Check the quickstart guide for more information.
Alternatively, a quick Google search shows there are various plugins/assets available directly for Unity3d which may be better suited.
I am working on web speech recognition.
And I found that Google provide a API which call "Google speech API V2" to developer. But I notice there is a limit on every day to use it.
After that I found there is a native WEB Speech API also can implement the speech recognition. And it just working on google chrome and opera:
http://caniuse.com/#feat=speech-recognition
So
1. What is the different Google Speech API and Web Speech API? Are they have any relations?
The speech recognition result json is return from google. Is that the google speech api will be more accurate than web speech api?
Thank you.
The Web Speech API is a W3C supported specification that allows browser vendors to supply a speech recognition engine of their choosing (be it local or cloud-based) that backs an API you can use directly from the browser without having to worry about API limits and the like. You could imagine that Apple might power this with Siri and Microsoft might power this with Cortana. Again, browser vendors could opt to use the built in dictation software in the operating system, but that doesn't seem to currently be the trend. If your trying to perform simple speech synthesis in a browser (e.g. voice commands), this is likely the best path to take, especially as adoption grows.
The Google Speech API is a cloud-based solution that allows you to use Google's speech software outside of a browser. It also provides broader language support and can transcribe longer audio files. If you have a 20min audio recording you want to transcribe, this would be the path to take. As of the time of this writing, Google charges $0.006 for every 15s recorded after the first hour for this service.
The Web API is REST based API with API key authentication, especially for web pages which needs a a simple feature set.
While Google Speech API basically is a gRPC API with various authentication method. There are lot feature is available when you use gRPC, like authentication, faster calling, and streaming!!!
I want to integrate voice detection in my iPhone app. The iPhone app allow the user to search the word by using their voice. But, i don't know a single info about Voice Recognition in iPhone. Can you please suggest me any ideas,tutorials or sample code for this?
You can also use Google Chrome API to integrate voice recognition on your application, but there is a big problem : the API works only with FLAC encoded files, but this encoding isn't supported natively on iOS... :/
You can see those 2 links for more information :
http://www.albertopasca.it/whiletrue/2011/09/objective-c-use-google-speech-iphone/
http://8byte8.com/blog/2012/07/voice-recognition-ios/
EDIT :
I realized an application including voice recognition using Nuance SDK, but it's not free to use. You can register for free and get a developer key that allows you to test your application for 90-days. An application example is included, you can see the code, it's very easy to implement.
Good luck :)
The best approach will probably be to:
Record the voice on the phone
Send the recording to a server that runs the speech recognition software
Then return something to the phone to indicate what it should do
This approach is favorable as there are a number of open source voice to text softwares out there & you are not limited by computing power in the backend.
Having said that, iOS has OpenEars which is based on Pocket Sphinx. It looks promising...
Well voice recognition is not correlated with iphone. All you can do is record the voice in iphone. Once done, you can either code your one voice recognition module, or find a third party API and reuse it.
You can do google search on that.