I am currently working with Alexa Smart Home Skills and implementing an interface to various data that I already have in a database. For example, this was no problem for the temperature, since there is a known API for this:
https://developer.amazon.com/en-US/docs/alexa/device-apis/alexa-temperaturesensor.html
But now I wonder why there is no API for humidity or brightness? Of course I could also use a PercentageController for this, but that would probably eliminate the pleasant voice support of Alexa? It's important to me that I don't need an unnecessary activation word, as would be the case with normal skills.
Is there any other way to query humidity or brightness via Alexa?
Thank you very much and best regards!
There is API for brightness : https://developer.amazon.com/en-US/docs/alexa/device-apis/alexa-brightnesscontroller.html
Could not find an API for humidity. You will have to go with percentage controller only.
Related
I need to simulate a fully functional AS/RS in warehousing. Moreover, I am a complete beginner in this field. Can some please let me know if I could get readymade simulation file? Or if not, please let me know how to learn to do it.
I have checked out the Anylogic website and it's tutorials (They are too lengthy).
fortunately for you, i have developed an AS/RS example that is a ready-made downloadable model for you, available at https://cloud.anylogic.com/model/1f5c7d1f-8782-40ac-957d-d3ba97bf6bf0?mode=SETTINGS
In general, when you want a model example, the first thing you should do is check the anylogic cloud, and if you are lucky the model is downloadable. Unfortunately, most people don't share
It really depends on what type of ASRS you are modeling (shuttle versus unit load) and level of detail you need. Do you need specific slotting and inventory tracking, or simplier black box delays with the assumption inventory is always available? The level of detail you need depends on questions you are asking, and should be addressed prior to starting development. If results from the model are critical and urgent, and you need anything more than simple black box delays, you should consider outsourcing to an experienced professional until you can get your AnyLogic skills up to speed.
This might sound like a general question but I am learning AI techniques/algorithms to implement the best solution for my website to reply to my users. I want to reply to my users as most of their questions over the email are almost the same so I am not sure which algorithm to use to implement my mailbot.
I have been studying AIML and apparently it uses best pattern matches and respond with the predefined template thus very less capacity so self learn. Other than this I have started studying basic NLP which would tokenize, segment, and then create an answer based on that. Maybe I need some advice on where to look.
Any suggestions will be appreciated.
Thanks.
I'd recommend Dialogflow for any chat bot applications. Not only does it make ml chat bots easily accessible and free (within a very generous free tier) but also allows you to leverage Google's models and infrastructure with very little development time and effort.
You can even get away with writing zero code, depending on your required integrations or backend.
Dialogflow also allows you to easily implement multiple languages and deploy to many platforms with little in terms of manual work.
It'll be a little more effort than keyword matching but will be a lot more versatile and you'll learn a lot more about intents, context and ml chat bots.
For you to deploy via email you'll want to integrate the API which should be very simple once you have a model configured.
I want to create an app for OSX which would work as an addon (displaying some overlaying information) to other app. Something like Poker Tracker for example - it shows extra information for poker games while playing on tables.
Just wondering is it possible using Swift? Can you point me to some direction what to look for? some libraries helping with such case? Never developed anything for OSX but keen to learn.
Thanks in advance.
Just wondering is it possible using Swift?
Yes, you can use Swift to create macOS applications. It's not magic, though -- your Swift code can only do things that are actually possible.
Can you point me to some direction what to look for?
Look for an API that lets other apps interact with the host application. That API will define what your "add on" application can reasonably do.
Without some sort of API or scripting interface, it's going to be very difficult to write a program that interacts with the host application. The best option is probably the Accessibility API in macOS. Accessibility is meant as an assistive technology, but it's often repurposed for tasks like automated testing. You might be able to use it to gain some level of control over the host app.
As far as I know it doesn't expose any API, so it would need to be image scraped.
This is really a tall order, and doubly so if you're asking basic questions about language capabilities. I think you'd have much better luck creating an efficient user interface so that the user could enter the relevant information directly, e.g. what cards the other users are showing, bet sizes, etc.
I use MapManager to show the route. MKroute doesn't work in my country, but it does in US. Can anyone tell me if this only work in only particular countries or it does work everywhere?
The best i can tell is that, for a route, you need a destination and starting point.
When a coordinate preferably the current user position does not have a POI in map mostly due to the fact that apple maps is not completely mapped in your place. By which i mean its besides from showing some main areas and streets, you have nothing other then this feature will not work.
If you know what i mean, then the alternative solution is Google Maps. It has a vast resource and it has information about things apple does not.
But it comes with a price, you need to learn to use Google Maps API which at this point is documented for Objective-C only.
And the callout is somewhat static in Google Maps which you can customize but the hard way.
I'm trying to create an iphone application that allows a user to take a photo of their smile, then drag and drop new smiles, from a small predefined list.
I know there are a lot of photo manipulation apps and I have seen similar concepts that allow smile manipulation, but not quite what I am looking for. The problem is knowing where to start. How can I create this effect? Would the OpenCV iPhone port be the best way to go? Or perhaps something using OpenGL? Willing to do some research, but I find that experience often goes a long way, so any advice or insight would be much appreciated.
That's a pretty cool application. I'd recommend some type of dense optical flow type alignment method which would be strike a balance between global consistency and local consistency.
In short, you'll want some generic mouth shapes in a gallery. Then, you can crop the user's mouth region and warp it to the gallery shape to show what their smile in that shape will look like.
Ce Liu's Optical Flow implementation might be an interesting point to start. You should be able to port that reasonably easily.