grafana alert dashboard UID - grafana

when I'm trying to create the alert and add details I see that it has a dropdown for Dasboard UID, which I assume will render a link in the notification to open a specific dashboard, however I have no idea what to put there.
I'm getting the following error:
Failed to save rule: invalid rule specification at index [0]: both annotations __dashboardUid__ and __panelId__ must be specified
so my dashboard URL is this
https://grafana.site.com/d/VkIxY9jnz/executerequest-error-log-analyzer?orgId=1&from=now-1h&to=now&refresh=10s
So what should become the dashboard ID and why do I need the panel ID? My dashboard has 2 panels - the chart based on loki agg query and below the raw loki data.

Related

Logic Apps trigger Databricks notebook

I have an issue. Is that possible for Logic Apps to trigger Databricks notebook in its pipeline?
I want to automate processing of downloaded files from Sharepoint and trigger notebook of Databricks.
Starting to build the custom logicapp connector but struggling according to this guide: https://medium.com/#poojaanilshinde/create-azure-logic-apps-custom-connector-for-azure-databricks-e51f4524ab27
Unfortunately struggling to create json Openapi file (image below)
1.Created a Databricks resource in Azure.
Created a Logic Apps Custom Connector resource with name Databricksconnector.
Once created clicked on edit option as shown below,
Followed the document https://medium.com/#poojaanilshinde/create-azure-logic-apps-custom-connector-for-azure-databricks-e51f4524ab27
by # Pooja Shinde
and performed below steps.
Connector creation I have kept all with default values in this section.
In General Information, selected scheme as Https and in host give data bricks url as https://databricksinstancename.azuredatabricks.net
In Security, selected authentication as basic as shown below,
In Definition tab, added new action by clicking on New action
In General tab, provided Summary, Description, Operation ID and Visibility selected as none.
In Request tab, Clicked on Import from sample.
I am retrieving job details by passing job_id and details are,
Request Type: Get
Url: api/2.0/jobs/get?job_id=*******
Headers: Content-Type: application/ json
12.Click on import and request will be update as below,
Next click on Update connector and connector will be saved.
Connector saved successfully as shown below,
We can view all the provided details in json format by clicking on swagger editor,
Now created a logic app with recurrence trigger.
Able to see created custom connector as shown below,
Created connection for Azure data bricks as per the document https://medium.com/#poojaanilshinde/create-azure-logic-apps-custom-connector-for-azure-databricks-e51f4524ab27
Added parameter “job_id” and given value as shown below,
20. logic app ran successfully and able to get job details,
Reference: https://learn.microsoft.com/en-us/azure/databricks/dev-tools/api/2.0/jobs?source=recommendations#--runs-list

Azure API Manager - Issue on URL , working even if no sense url and just getting the last segment

We have found this bug in Api Manager that whenever you pass some similar words in the url its working properly even the url was not correct .
Example:
working url: {{Url}}/regulators
bad url but works with the same : {{Url}}/../auth/streams/../../regulators
Anybody encounter the same issue with your APIs?
Please see image below
Method GET is used to add form data to the URL in name or value pair. If you use GET, the length of URL will remain limited. It helps users to submit the bookmark the result. GET is better for the data which does not require any security or having images or word documents.
Also, this is used only to get the data from address bar in the browser.
We can check with below settings for API Responses:
Select the API you created in the previous step.
Select + Add Operation.
In the Frontend window, enter the following values.
Display Name(test call) : The name that is displayed in the developer portal.
URL(HTTP Verb) – GET : Select one of the predefined HTTP verbs.
URL (/test) : A URL path for API.
Description: Optional description of the operation, used to provide documentation in the developer portal to the developers using this API.
Select the Responses tab, located under the URL, Display name, and Description fields. Enter settings on this tab to define response status codes, content types, examples, and schemas.
Select + Add response, and select 200 OK from the list.
Under the Representations heading on the right, select + Add representation.
Enter application/json into the search box and select the application/json content type.
In the Sample text box, enter { "sampleField" : "test" }.
Select Save.
Refer to MS Docs to Monitor published API’s

Google Data Studio error - User Configuration Error This data source was improperly configured. Error ID: fb2ea24f

I am trying to create a few visualizations from BigQuery data on Google Data Studio.
Following are the simulation steps:
Login into the BigQuery console
After running a query on the new Bigquery console, click the option "EXPLORE IN DATA STUDIO".
Create any chart or table
Click "SAVE" on the top right corner
This will generate an error:
User Configuration Error This data source was improperly configured. Error ID: fb2ea24f
Page with default table after clicking "EXPLORE IN DATA STUDIO":
Table disappears after clicking "SAVE":
Error details shown on pop-up after clicking "See Details":

SuiteCommerce Advance: Unable to get the Free Shipping Rule Rate 'Insufficient permissions' for Custom Service

We tried loading shipping item data with 'nlapiLoadRecord', Currently we have two methods available Standard & Express. We tried to load data for standard method as below in Script Debugger.
​var standard = nlapiLoadRecord('shipitem',3);
We are getting all fields of standard shipitem in Script Debugger, then we are fetching shipping rule value with
var standard_amount = standard.getFieldValues('freeifordertotalisoveramount');
Then we tried to load these same data using nlapiLoadRecord in our custom SuiteScript Model, we tried to call this method in our custom ShippingDetails.Model, and tried to fetched the data on cart page but it is giving us 'Insufficient permissions' error massage. Please see below attached screenshot. nlapiLoadRecord('shipitem',3); does not allow us to load data on front end.
Basically this is not case with your code, as you are trying to load shipping rate using SuiteScript Service and model this is perfect.
​var standard = nlapiLoadRecord('shipitem',3);
var standard_amount = standard.getFieldValues('freeifordertotalisoveramount');
This is completely related with user's permission i.e Roles Under Setup-->User/Roles-->Manage Roles you are trying to fetch data of 'shippingitem' in this case you must need to check different permission
Check your Script Service is enabled, enabled for with out login in NS back end
Most Imp:
2. Check shipping item is added in permission under List i.e the role which you are using for Service, for that role you must need to set shipping item See below screen shot you will get more idea.

Error when trying to load tables from desktop with dashDB

I'm trying to load a csv file into a table and receiving the following error:
The steps that I am doing are launching the dashDB instance that I have, Load, Load from Desktop, Browse for the csv... I have tried to create a new table from the csv and add to an existing table, both resulting with the error. I have also created a dummy csv with just 1 record in it and it is also failing with the same error. When creating the table from the load, the result is the table getting created with no data loaded.
The error you are getting is an already known bug of the Bluemix Dashboard integration with the dashDB one:
The team is working on this issue, but a fully working workaround is available.
In order to complete your data loading through your file you should:
access your application section on the Bluemix dashboard
click on the 'Environment variables' section and select the 'VCAP_SERVICE' button
retrieve the 'https_url' of the dashDB service and save also the related username and password values
paste on your browser url the https_url value, it will open directly the dashDB dashboard, you can now login using the username and password found in VCAP_SERVICE section
click on 'Load' section in the left menu of the dashDB dashboard
in this section you could upload your file without any problem.