Upgrade Grafana 7.x.x to 9.x.x alerts missing - grafana

I had Grafana 7.5.7 working with a lot of alerts configured I used to save all the dashboards as a json file and this has the alerts. After I upgraded to the 9.3.1 version all the alerts disappeared.
I saw online that you can configure the old alerts
force_migration = true
[alerting]
enabled = true
[unified_alerting]
enabled = false
But his does not work. Can someone give me some insight on what I need to do? Should I just create new rules and provision them ou is there a was to migrate these?
Thanks in advance

Related

bitnami.keycloak helm chart - how to turn of theme caching

I'm trying to get the Bitnami Keycloak instance to start in dev-mode and/or with the theme cache turned off. I've tried
extraEnvVars:
-name: KEYCLOAK-PRODUCTION
value: false
and also
cache:
enabled: false
I've tried using extraStartupArgs setting the spi values as suggested in the Keycloak docs. https://www.keycloak.org/docs/latest/server_development/#creating-a-theme
Nothing is working. I'm currently having to update the CSS file name with a version number and also updating the theme.properties file with that version number too and I'd like Keycloak to just reload everything with each page load.
Hope someone can help out.
Scott

Grafana on ECS cluster return not allowed origin

Created ECS cluster let's call it tools.
tools cluster have 2 services:
sso-proxy
grafana - opensource
Going through sso-proxy to get grafana.
when trying to enter credentials to grafana i got the following error:
When incognito mode the login working for the first time.
This is the ECS logs:
Please someone can help.
Grafana must receive proper Host header in the request header. I guess your "sso-proxy" doesn't do that. You didn't provide reproducible example (why not, when you want to help), so it is only a guess.
Lazy and insecure workaround will be Grafana downgrade to version 8.3.4-, where CSRF fix for CVE-2022-21703 is not included.

How can I set grafana 7 to work with warp10 2.7.x?

I've just install a fresh warp10 standalone server version 2.7.2.
Using beamium to send data into it and seeing the data via the VScode plugin is OK and I can see graph in GTS preview.
I've also installed a fresh grafana and warp10-plugin following the warp10 recommandation on the ovh github.
When execution a default warp10 query (via editor), grafana add some strings in the query, like the start or end value, so in the end the query look like:
1609947849757000 'start' STORE
'2021-01-06T15:44:09.757Z' 'startISO' STORE
1610034249757000 'end' STORE
'2021-01-07T15:44:09.757Z' 'endISO' STORE
86400000000 'interval' STORE
199538106 '__interval' STORE
199538 '__interval_ms' STORE
[ 'host.local.domain' ] 'host_list' STORE
But when executing, there is an error poping in the warp10 logs, after decryption it tell:
Exception at '=>1609947849757000<= 'start' STORE '2021-01-06T15:44:09.757Z'' in section [TOP]
It seems that it don't take the LONG (DOUBLE?) number for what it is, and try to match a function with the same name that don't exists.
On grafana side, I don't have any valuable help, it tell me:
WarpScript Failure on Line -1, Unable to read x-warp10-error-line and x-warp10-error-line headers in server answer
Do I miss something ?
==== Edit: 2021-01-07 17:32 UTC
After first reply, doing other test:
I've tryied the same query, and the error is still the same.
Warpscript failure
But in VScode this query works:
{
'token' $RTOKEN
'class' '~.*'
'labels' {}
'end' '2021-01-07T17:35:28.086Z'
'timespan' 21600000000
} FETCH
I've also try to use the bartender stuff in grafana and it work fine too...
So everything should work, I must miss the obvious.
Can Java version have an impact ?
If the datasource is working, you didn't miss anything.
Do you use the built in warpscript editor ? Make sure you ticked the "WarpScript Editor" checkbox:
Then, do the simplest FETCH request you can do, or copy paste the code you did in VSCode.
You can define your own variables too in the datasource configuration. It is usefull for tokens.
I just setup Grafana 7.3.6 + ovh plugin to check, it seems there is no regression issue.
The full WarpScript can be found in the browser console.
The header error is linked to Grafana 7, no problem with Grafana 6.x. Install Grafana 6.x if you want WarpSript detailed errors.
Edit: Issue is now fixed and merged into ovh github master.
If you want to test another Warp 10 data source, you can use the credential and advices in this article. You just need to go before covid lockdown to find relevant office beertender data...

IBM Cloud Private CE - Unauthorized Access to Catalog

I have installed ICP CE 2.1.0 on a google cloud VM and the installation has gone well-no errors in installation process. When accessing the GUI I am able to see deployments and services but as soon as I access any part of the Catalog I get a blank white page with the text:
{"statusCode":401,"details":"Unexpected response code 401 from request:\nGET https://xx.xxx.xxx.xx:8443/console/api/v1/header?serviceId=catalog-ui&dev=false&accessUrl=https%3A%2F%2Fxx.xxx.xxx.xx%3A8443* ...... }
I have tried killing the individual pods but I get same error. When looking a the pod logs for the catalog-ui I have error 500 messages.
Has anyone experienced this or can tell my why this is the case? Understand that a cloud VM is not the best use case maybe but it should work?
Can you confirm the version level of ICP? Your post mentioned "ICP CE 2.1.0" but if you can check the user icon (top right corner) and click About, we should be able to see the full version details.
Reason for asking is that, at the 2100 level there was an intermittent catalog issue just like you describe. Generally it was caused by resource constraints on the k8s.
Details for ICP 2103, which is the latest available release:
https://www.ibm.com/support/knowledgecenter/SSBS6K_2.1.0.3/getting_started/whats_new.html

Installing openTSDB on Ubuntu15.04

I have installed openTSDB(.deb package) on ubuntu 15.04 by following the guidelines stated in documentation. when I give this command "service opentsdb start" it is not starting and it is mentioned in documentation that we have to change some configuration files.can anyone please tell me what are the changes that we have to do and in which file the changes have to be done?
Thanks in advance
Regards
VHC
Check your logs in /var/log/opentsdb/opentsdb.log. You should have HBase up&running correctly (means i.e. you are able to create table and store some values)
Remember you have to create tables in HBase running
env COMPRESSION=NONE HBASE_HOME=path/to/hbase-X.XX.X /usr/share/opentsdb/tools/create_table.sh
http://opentsdb.net/docs/build/html/installation.html#create-tables