(I asked this also on Datalore's forum. There doesn't seem to be much going on there -- so I'm asking here in the hope of a better/quicker response.)
I'm having trouble attaching GCS buckets. The documentation is sparse. All that I could find is here, which simply says:
In the New datasource dialog, fill in the fields and click Save and close.
Here's that dialog and I'm not sure what information to put here.
What I tried
GCS datasource name
I believe it's for reference within Datalore, correct? So can I just put anything here? I wrote "patant-data-ors".
bucket
Options I tried:
patent-data-ors (this is the name of the bucket)
storage.googleapis.com/patent-data-ors
patent-data-ors.storage.googleapis.com
Also tried 2 and 3 with https://.
key_file_content
I left it blank. I'm guessing it's for private buckets? Mine is publicly accessible.
What am I doing wrong?
I have a task where I need to use ARM template to "setup SQL server with elastic pool and 2 databases (imported from bacpac files) use Microsoft Adventureworks samples."
Most of the template I made using "101-sql-elastic-pool-create" template that is available here: https://github.com/Azure/azure-quickstart-templates/tree/master/101-sql-elastic-pool-create
I used it for the elastic pool setup and used a sample from this documentation (https://learn.microsoft.com/en-us/azure/azure-resource-manager/templates/template-tutorial-deploy-sql-extensions-bacpac) to have the .bacpac imported to the database
But when I try to deploy this template (the template is available to see here: https://github.com/KarlisAG/SQL_ARM/blob/main/azuredeploy.json ) it throws me an error
("Resource Microsoft.Sql/servers/elasticPools 'task4sql/elasticp' failed with message '{ "code": "BadRequest", "message": "An error occurred while processing this request.", "target": null, "details": [], "innererror": [] }' ")
So, the biggest problem is (besides me being an absolute newbie) not knowing what I need to fix, because there is no additional information granted, besides the "BadRequest". For deployment I used two different methods: 1) the automatically generated one from VS when created Azure resource groups, 2) One shown in this documentation, with few of my own changes, so it could work for me ( https://learn.microsoft.com/en-us/azure/azure-resource-manager/templates/template-tutorial-deploy-sql-extensions-bacpac#deploy-the-template ), from where I also got the code part for importing .bacpac to my newly created database. [Both of the are available at the same repository, named "Deploy-AzureResourceGroups.ps1" and "DeployScript.ps1" respectively].
Both of the methods return the same error.
I even manually created these things in Azure portal to see what their template looks like and maybe copy some things that are not present in my template, but from what I saw I have exactly what they automatically generate.
And when I checked the deployment status in Azure portal I can also see that during the elastic pool creation it fails: screenshot from that section available here
So does anyone know how should I have created this type of template, or did I recreate it poorly, or am I missing some small mistake that keeps blocking me?
So after multiple tries it seems that I finally figured out what I was doing wrong. I am still not sure which of the changes made a difference, besides the last one, so I will write what I changed, so other people who could be facing similar issue know what they could do fix the issue.
At first, after #stringfellow noted, I thought that Free Trial is restricting me from doing anything, but after I lowered my DTU multiple times, to the point where it was at 5 (starting from 100), the same issue persisted, so probably it was not the issue (later on I had to change it at minimum to 50). I also changed the Elastic Pool edition from Standard to Basic. But doing these changes still didn't fix the issue.
At last, after answering #NachoMartínez-Aedo comment about the api version, it got me thinking. Because in the code that I posted I used "2020-08-01-preview" api version, but in reality I used "2014-04-01-preview", because when I changed the api version of all my parts to the 2020 one, then powershell told me that (No registered resource provider found for location 'northeurope' and API version '2020-08-01-preview' for type 'servers/databases/extensions'. The supported api-versions are '2014-01-01, 2014-04-01, 2014-04-01-preview'.). And after reading through that again and remembering that when I was looking at some documentation and Azure portals automatically generated templates I knew that I always say them using the 2020 api, not the 2014, which left me confused as to why I couldn't also use the newest version. And reading that error again I saw, that it specifically told that 2020 api version was not supported for the type "servers/databases/extensions", not that all of the parts must be in 2014 version (which is what I thought when I saw that error), so I tried and changed all the 2014 back to 2020 and left the extension api back at 2014 (previously I thought that all api versions should be the same, not sure why, maybe because most of the examples and documentation had everything have the same api version) and the everything worked, even the import part, about which I was worried.
So in the future I am not too sure if there is a better/newer way to import .bacpac file to newly created database with ARM, or just have to leave it as is and in the future use this same approach. Does anyone know?
I got the import extension from here: https://learn.microsoft.com/en-us/azure/azure-resource-manager/templates/template-tutorial-deploy-sql-extensions-bacpac
, where they use examples with 2014 api for the extension, so because of that I tried to use it for all my parts, but only during this answer writing did I realize and look through it again, that they used the 2014 api only for the extension part and not other parts, like server, database, etc.
TL;DR
So basically I needed to use the '2020-08-01-preview' api for everything besides from the extension part, where the newest api version is only '2014-04-01-preview'
I was able to deploy most of your template and use similar parameters in my own subscription. I saw in your screen capture that you are using a Free Trial subscription. I am curious if that has anything to do with the 'BadRequest' due to any sort of DTU limitation.
Here is my deployment, which failed as expected when it tried to import the databases.
I created the watson service and went to create a new project. It expects me to add Cloud Object Storage. When I chose the Lite option and press create I get this error "Reseller channel 2c95500b-ea86-4b13-8bb5-b2f0c2fa8200 is invalid."
My goal here is to setup the Watson service in order to play with Visual Recognition dk service. But I am stuck at this stage.
Can someone help with some suggestion of what else can I try or what I might be doing wrong?
Thanks!
I also struggled with this a few minutes ago.
And Cloud Object Storage link at the bottom of the page under Choose project options won't help,
but i found another solution that works for me
Go to https://www.ibm.com/cloud/object-storage/pricing
and click Order IBM Cloud Object Storage at bottom of the page
I hope this can solve your problem
I am starting from the position of near absolute ignorance here.
I successfully set up a Bluemix account, added devices, got API keys, connected my Raspberry Pi running Node Red and a temperature sensor to Node Red on Bluemix, created a dashboard and was able to get data back and forth between Watson/Bluemix and my Pi. Cool. Unfortunately, I got all that way without ever really understanding the navigation of Watson/Bluemix nor all the terminology. I made NR work both ways and was happy.
Then one day last week, the route to my Bluemix Node Red application stopped being accessible:
404 Not Found: Requested route ('misterpi-01.mybluemix.net') does not exist.
I can browse my devices and see that they are sending data every few minutes, so clearly the connection is there. When I click on IBM Cloud Apps, I am told
There are no IBM Cloud applications bound to your organization
When I am on my IBM Cloud Dashboard and I click on Connections I see I have a connected application that is "Not Running" with a red dot. I can click Stop and the status goes to Stopped. When I click Start, after a short while,
I cannot add an another app because I can only have one.
Where did my NR app go and can I get it back (preferred) and if not, delete it?
At the suggestion of #hardillb I went looking for logs. I found them and it looks like the app is continuously started and crashing. I can't copy the text from the error logs but here is a screenshot.
The first thing you should do is get a copy of your flow source for backup. Go into the IBM Cloud Dashboard and click on Cloudant-ll (the suffix may be different?) and then Launch Cloudant Dashboard (alternatively, click on ...-cloudantNoSQLDB and then alias of Cloudant-ll). From there click on NodeRed and then select the id that ends with "/flow". Then you can extract everything inside "flow": [].
I would recommend emptying it out so you have no flows to ensure that Bluemix can start 'empty'. If you do this you will end up with something similar to mine below (don't change _rev).
{
"_id": "Home Assistant/flow",
"_rev": "102-2b38e1756eeb26def19b20858bc50a40",
"flow": []
}
Later you can paste the extracted code back into Node Red using Import > Clipboard.
You may also want to edit the ".../settings" document to remove any extra node types you installed.
Caveat - I had the same problems with Bluemix and eventually gave up - crashing at least weekly running out of memory and difficulty restarting. I suspect it can't reliably run with 256MB.
I had similar issues with log/errors and endpoint 404:
I followed Michaels advice to backup (nodered:flows and settings) in cloudant.
1. However what works(for now) was cleaning part of the nodered /settings.
2. Basically I removed the nodes part from the Json up until almost the end, leave _credentialSecret part at the end.
3. restart and go to your url, you will see your flows
4. if your flows use some nodes that produce warnings, it might be necessary to remove them to avoid future problems.
UPDATE 2: I saved every flow separately, and then deleted them all from the nodered flows. Then restarted, and have been importing one by one, and it is working.
I am using the web interface, I am trying to make public files that were uploaded and each time I click the checkbox I immediately get the error "Invalid Argument" and the box unchecks.
This also happens while trying to uncheck files already uploaded and public.
I'm not certain if something happened to my account or if this is a service problem for the site but we're in a major jame because of this. Essentially Cloud storage is unusable at this stage.
There was an issue with the web user interface and the JSON interface affecting setting ACLs on objects. That problem has been resolved, and all interfaces should be working correctly now.