In the readme it says
If you are only interested in building Dialogflow fulfillment for the Google Assistant and don't plan on using other integrations, please use the Actions on Google NPM module (actions-on-google) which supports all Actions on Google features.
Could anyone provide some info on what features we would miss out on if we use dialogflow-fulfillment-nodejs instead of the actions-on-google one?
There are many features that are part of the Actions on Google platform which are not available more generally to other platforms, necessitating the creation of a separate library.
Transactions
User storage
Daily updates and notifications
Quick Google Sign-In
Permissions to get things like user location
Deep-linking into an Android app
The ability to check surface capabilities and move the conversation to a new surface
And many more advanced features
Related
I am currently developing an updated version of a voice app that it was previously developed with Google Actions console as a conversational app. However, when started to create a new project for this updated app, a banner says that Conversational apps will be sunsetted on June 13th.
Reading through the documentation, it is not clear to me how to develop now that this options will not be available in the future. Among the options, App Actions and Dialogflow CX might be the route for development. However, my app to be developed reqquires integration with smart devices such as Google nest Hub and Google nest mini for interaction.
It seems that App Actions might not be the solution because it adds voice capabilities to an existing Android App, which I am not sure that will work with Smart devices directly.
On the other hand, dialogflow CX looks to be focused on chatbox, i.e. based on text. Again, I am not sure if using Dialogflow CX will be able to provide an appp that uses voice interactions in smart devices.
In addition to these options, I also read about Dialogflow ES, Cloud-to-Cloud for smart for Google Home, Content Actions, and Media Actions. It does not look like either of these solutions might be a replacement to Google Actions. For example, I think smart home is not the option because I am not looking to interact with home devices besides google nest hub.
I hope anybody can help and help me understand development path or lead me to information that I might be missing. Thank you
It sounds like you are looking to create a new version of a conversational app that was previously developed using the Google Actions console, but are now facing the sunset of the conversational apps feature on June 13th.
One option that you may want to consider is Dialogflow CX, which is a platform for building and managing natural language conversational experiences. Dialogflow CX is built on top of Dialogflow ES, and is designed for more complex and large-scale conversational apps.
It allows you to create a conversation flow using a visual editor, and supports integration with various platforms including Google Assistant, Alexa, and WhatsApp.
Regarding the integration with smart devices like Google Nest Hub and Google Nest Mini, you can use the Google Assistant Actions API to build custom actions that can be invoked by users through the Google Assistant on these devices. This API allows you to define a conversation flow and handle user inputs, and it can be integrated with Dialogflow CX to handle natural language understanding and generation.
App Actions, on the other hand, are a way to surface your app's functionality through the Google Assistant on Android devices, and it's not a replacement for Google Actions.
Cloud-to-Cloud for smart home, Content Actions, and Media Actions are more specific solutions for different use cases, for example, providing actions for smart home devices, providing actions for media content and etc.
In summary, Dialogflow CX along with the Google Assistant Actions API seems to be the most relevant solution for your use case. It allows you to build natural language conversational experiences that can be integrated with smart devices like Google Nest Hub and Google Nest Mini.
It's worth checking the official documentation and tutorials for more information about Dialogflow CX, the Google Assistant Actions API, and how to integrate them. Also, you may want to reach out to Google support team for more specific guidance for your case.
I'm very new to google actions and I got one project which is made by using conversational actions. (You can try it in google assistant talk to help me sleep)
And as you all know conversational actions will be deprecated soon, and I try a lot for alternatives for Google Home Nest Mini but I'm stuck and I can not force my end user to install any android application.
Currently we do not provide API for configuring the smart home actions.
You can open a ticket in this regards on the public issue tracker and we will forward your request to the internal team to see what we can do in this regard.
Marketing specialists keep instructing my client to setup the analytics so that it can track where the user is being referred from so that we can optimize the advertising campaigns.
When I ask them how to setup the tracking correctly they send me to a link for the native Android & iOS google analytics libraries and provide the settings. I respond in kind stating that our app is not built using the native SDK's it is built using the Ionic Framework. They then have no response for this.
I have Google Analytics implemented as well as Facebook analytics. Both systems are actively tracking data. But apparently it is still not tracking the source of which advertising campaign the user originated from and / or installed the app from.
I've tried to find an example of this somewhere and talked to numerous people but from what I can tell it is not possible using the Ionic Framework.
Can anyone provide some insight into this? If this is not possible it seems like a rather large issue with the Ionic Platform that should be made more visible.
Thank you!
I am new to google home. I had an experience in Amazon Alexa custom skill development. In Alexa, I had deployed my codebase in lambda function and also test my custom skill using actual Alexa device register with my email id.
So, Now I need to develop similar skill in google home device. Till now, I didn't get any good tutorials.
Is it possible to create & test Google Home app like Alexa skill?
The steps you go through to develop a Google Home action or app are very similar to creating an Alexa skill. There are a couple of differences, but logically they are the same.
If you use a NLP system such as Dialogflow (which is strongly suggested), you build the suggested phrases that the system responds to and the Intents they correspond to. You would specify your webhook as part of building these phrases. If you don't wish to use an NLP, you can specify the initial Intent phrases using the Actions SDK and specifying the configuration as part of a .json file. Other actions you'd do in the Alexa console are similar to what you'd do in the Assistant console.
You can deploy your Action on any public server that accepts HTTPS connections. This can include AWS Lambda with an AWS API Gateway trigger, or a Firebase Function, or a web server you more directly control that has a valid SSL certificate. This webhook would get a JSON body and needs to send back a valid JSON response. Google has libraries for node.js to help with this.
Google has a relatively full-featured simulator which you can use to test your Action. Once it is available in the simulator, it is also immediately available on every device attached to that account. You can permit other accounts to the project as well and, once they have activated it through the simulator, it is available on all their devices as well.
A full set of documentation is available at https://developers.google.com/actions/. It includes links to sample code, and you can find more step-by-step codelabs at https://codelabs.developers.google.com/?cat=Assistant
If you're familiar with how to develop skills for Alexa you might want to check out the jovo-framework. It makes it pretty easy to create skills that work for both Amazon Alexa and Google Home.
Here is a good starter template and walk-through that will get you going. https://github.com/rmtuckerphx/ask-cli-jovo-starter
Do you think that if I'll build my own custom analytic tool (Flurry, Pinchmedia) and I'll host that on the same server where I have my data source for the application, will Apple consider this as a thirdparty analytic tool or not? ... Problem is that Flurry and Pinch are being banned from Appstore by the newest T&C ... than I thought that I'll build an open source library that will allow anyone to have their own analytic installed on their server ...
Thanks,
Ondrej
(Full disclosure: I work for Localytics)
What you describe would seem to comply with Apple's new terms. It's what I call 1st-party collection of device data. Or Apple could approve your use 3rd-party analytics, which is more likely if your analytics service isn't using those data to serve ads or selling data to someone else.
But why build it yourself? Localytics announced its Enterprise analytics service last week, which includes 1st-party data collection: http://www.localytics.com/blog. Localytics client libraries for iPhone (and Android and BlackBerry) are already open source.