Which is better for building chatbot - chatbot

I need to build a chatbot . so should I build one in Python or should I use dialogueFlow
I tried building with DialogFLow , but at times I get stuck in identifying the langauge of the user

There are number of chatbot platforms and NLP engines from which you can try from.
Few of them are : Dialogflow, Amazon Lex, LUIS, Chatfuel, Botsify, Beep-boop, Motion.ai, QnA maker, Recast.ai, Octane.ai etc.
It all comes to your requirement and trying out what suits you better.

Related

How do I make chatbot for my e-commerce site?

I want to do chatbot for e-commerce site.I couldn't find much resources to guide me.Can you help with how to create chatbot using rasa? or is there another method you suggest?
if you have sensitive data in your chat messages, you should go with rasa, since rasa will run on your own servers. but, if your data is not that important, you can use Dalogflow (a google product), which you do not need any programming knowledge. the best place to learn rasa is their official documentation.

How to use json in intent functions in dialogflow

I'm making an application for google home as graduation work from my university of systems analysis and development. My difficulty is integrating with an external server.
I'm doing the code in javaScript, node.js and my intents call data from an external service, the server of the company, there is authentication, but even with Json without validation I can not access the data.
I used some examples of the google channel itself in youtube and the documentation of DialogFlow but I can not.
Has someone already done something similar and can you help me?

How to deploy and maintain multiple chatbots agents

This post follow this one where I explain one of my problems. Currently, I have to found a way to publish and maintain a high number of agents. I am not limited to Dialogflow.
I need some integrations like the google assistant (text and vocal), facebook messenger, telegram and if possible others like Slack, Twitter, Twillio, Alexa...
Okay, so I have already produced some agents with Dialogflow to understand the technology. I also read some pages of the actions-on-google documentation and I did'nt found anything on this subject. So basically I have to implement this:
Deploy around X agents through differents integrations instanciations. I mean I really need X facebook contacts, X google assistant apps, etc.
Maintain one code-base but have the ability to add localized-features like the name of the chatbot, currency or just block some intents (for Dialogflow example but in a more generic way, dialogs triggers).
It is just possible ? I am thinking about a web UI that can handle some facilities like the deployment, the monitoring and the maintenance. I am wondering if it's not overkill and if a more easier solution than mine exists already.
It isn't currently possible to create agents automatically, although Dialogflow's V2 API provides a mechanism to update agents via JSON once they have been created; see the restore and import endpoints.

Do I have to use API.AI to create an action for Google Home?

I have some experience building chat and voice agents for other platforms, but I’m not using API.AI to understand natural language and parse intents. Do I have to replace my existing solution with API.AI?
Not at all. The advantages of using API.AI in creating a Conversation Action include Natural Language Understanding and grammar expansion, form filling, intent matching, and more.
That said, the Actions on Google platform includes a CLI, client library, and Web Simulator, all of which can be used to develop an Action entirely independent of API.AI. To do this you’ll need to build your own Action Package, which describes your Action and expected user grammars, and an endpoint to serve Assistant’s requests and provide responses to your users queries. The CLI can be used to deploy your Action Package directly to Google, and you can host your endpoint on any hosting service you wish. Google recommends App Engine on Google Cloud Platform.
I found this explanation from the official page most helpful.
API.AI
Use this option for most use cases. Understanding and parsing natural, human language is a very hard task, and API.AI does all that for you. API.AI also wraps the functionality of the Actions SDK into an easy-to-use web IDE that has conveniences such as generating and deploys action packages for you.
It also lets you build conversational experiences once and deploy to many other platforms other than Actions on Google.
ACTIONS SDK
Use this option if you have simple actions that have very short conversations with limited user input variability. These type of actions typically don't require robust language understanding and typically accomplish one quick use case.
In addition, if you already have an NLU that you want to use and just want to receive raw text and pass it to your own NLU, you will also need to use the Actions SDK.
Finally, the Actions SDK doesn't provide modern conveniences of an IDE, so you have to manually create action packages with a text editor and deploy them to your Google Developer project with a command-line utility.
Google is pushing aggressively everybody to API.AI. The only SDK they have (Node.js) no longer supports expected events for instance. Of course, you don't need to rely on their SDK (you can talk to the API directly) but they may change the API too. So proceed with caution.

Watson conversation - How to deploy?

I've created and trained a basic dialog and it's now ready to be used in my web site.
I can't found any docs to deploy and use the application.
Can anyone help me ?
You need to build a front end application that allows users to interact with the conversation service.
There are some generated SDK's for various programming languages that can help you in doing this.