where should HADOOP_HOME variable point to on BigInsights on Cloud? - ibm-cloud

I'd like to use the kite-dataset tool on BigInsights on cloud, and have installed it on the mastermanager node in my home folder as per the installation instructions:
curl http://central.maven.org/maven2/org/kitesdk/kite-tools/1.0.0/kite-tools-1.0.0-binary.jar -o kite-dataset
chmod +x kite-dataset
However, when I try to run it, I get:
$ ./kite-dataset
ERROR: Cannot find Hadoop installation!
You can fix this warning by setting HADOOP_HOME
Where should this variable point to on BigInsights on Cloud?

Not sure if this is helpful, but this page shows where HADOOP_HOME can typically be found on BigInsights 3 and 4:
For IBM InfoSphere BigInsights 3.x:
Set HADOOP_HOME to BigInsights_Install_Directory/IHC.
For example, /opt/ibm/biginsights/IHC.
For IBM InfoSphere BigInsights 4.x:
Set HADOOP_HOME to BigInsights_Install_Directory/hadoop.
For example, /usr/iop/4.0.0.0/hadoop.

Related

Unable to start confluent local services

I'm using http://packages.confluent.io/archive/7.3/ > confluent-7.3.0.zip
confluent local services start
getting the error below even though the file is there. Please help!
Using CONFLUENT_CURRENT: C:\Users\user\AppData\Local\Temp\confluent.727246
Starting ZooKeeper
Error: exec: "D:\\Development\\tools\\confluent\\bin\\zookeeper-server-start": file does not exist
bin\zookeeper-server-start.sh script is intended for Unix execution.
Confluent Platform does not support Windows
If you use WSL2, it may work there. How to install Kafka on Windows?

gcloud SDK: Unable to write file

I installed gcloud SDK with brew cask install google-cloud-sdk
$ gcloud container clusters get-credentials my-gke-cluster --region europe-west4-c
Fetching cluster endpoint and auth data.
ERROR: (gcloud.container.clusters.get-credentials)
Unable to write file [/Users/xxxxx/my-repo]: [Errno 21] Is a directory: '/Users/xxxxx/my-repo'
Now all permissions of the folder and recursive files are restricted to 600 (drw-------). Tried to reinstall gcloud but with no effect on its behavior.
I assume you're using macOS and I'm unfamiliar with it.
The gcloud container clusters get-credentials command should write to a file called ${HOME}/.kube/config.
The error suggests that it's trying to write the credentials to /Users/xxxxx/my-repo and this is determined by the value of ${KUBECONFIG}. Have you changed either ${KUBECONFIG} or your ${HOME} environment variable? You should be able to printf "HOME=${HOME}\nKUBECONFIG=${KUBECONFIG}" to inspect these.
You may be able to choose a different destination by adjust the value of KUBECONFIG. Perhaps set this to /Users/xxxxx and try the command again.
Ultimately, this is some sugar to update the local configuration file. It should be possible to create this manually if needs be. If the above don't work, I can update this answer with more details.

Unable to run Grafana server on local system

I have to create a plugin for Grafana. For that i cloned the code from Github and followed this. I have managed to build the grafana but having issue in running the server. I am getting the error
Grafana-server Init Failed: Could not find config defaults, make sure homepath command line parameter is set or working directory is homepath
I did find this issue at issue. I have copied the "sample.ini" and re-named it to "custom.ini" and used the following commands in the bin folder made after building the grafana
.\grafana-server.exe --config D:\Go-UpdatedWorkSpace\grafana\conf\custom.ini
.\grafana-server.exe D:\Go-UpdatedWorkSpace\grafana\conf\custom.ini
.\grafana-server.exe -config D:\Go-UpdatedWorkSpace\grafana\conf\custom.ini
.\grafana-server.exe -config "D:\Go-UpdatedWorkSpace\grafana\conf\custom.ini"
.\grafana-server.exe --config "D:\Go-UpdatedWorkSpace\grafana\conf\custom.ini"
.\grafana-server.exe "D:\Go-UpdatedWorkSpace\grafana\conf\custom.ini"
But non of them worked. I am still getting the error
Grafana-server Init Failed: Could not find config defaults, make sure homepath command line parameter is set or working directory is homepath
What am i doing wrong ? I am new to Grafana development, any help would be much appreciated.
I have managed to resolve the issue. I ran the Grafana locally by using following command
./grafana-server.exe -config "D:\Go-UpdatedWorkSpace\grafana\conf\defaults.ini" -homepath "D:\Go-UpdatedWorkSpace\grafana".
I executed this command on the "bin" folder created after building the grafana with GoLang. (There is also a bin folder inside the homepath directory, i think we can use any of then "bin" folder). Issue was that i was not providing "homepath" in the command. "homepath" is the directory where grafana code is cloned.
In case someone looking for Mac, the below works. I tried with grafana version 8.x . For default install, No need to give -config as -homepath should take care. this could run from any location.
/grafana-install-path/bin/grafana-server -homepath /grafana-install-path web

Confluent 5.3.0 - missing confluent shell script

I am trying to start Confluent 5.3.0 on Ubuntu and need some help.
The first thing I tried was navigating to confluent-5.3.0/bin and running sudo ./confluent start (because that's the way I've been starting Confluent 5.2.1).
This gave an error message: sudo: ./confluent: command not found.
I then checked the folder confluent-5.3.0/bin and found it's missing a shell script named confluent, which is included in 5.2.1.
I checked the Release Notes and the Quickstart. Both of them say I should start Confluent with confluent local start instead of confluent start. (See screenshot at bottom). However, I still get the same error message because the shell script is still missing: sudo: ./confluent: command not found
For those of you who are able to start Confluent 5.3.0, how are you doing it? Can you paste your command line here?
Did you have to write your own confluent shell script for 5.3.0?
Can I just copy my confluent shell script from 5.2.1? I'm assuming it must be incompatible, since they removed it from 5.3.0.
From 5.3.0 you will have to install CLI separately
https://docs.confluent.io/current/cli/installing.html#cli-install
curl -L https://cnfl.io/cli | sh -s -- -b /path-to-directory/bin
Then you can run
confluent local start --path <path-to-directory>

How to start up a Kubernetes cluster using Rocket?

I'm using a Chromebook Pixel 2, and it's easier to get Rocket working than Docker. I recently installed Rocket 1.1 into /usr/local/bin, and have a clone of the Kubernetes GitHub repo.
When I try to use ./hack/local-up-cluster.sh to start a cluster, it eventually fails with this message:
Failed to successfully run 'docker ps', please verify that docker is installed and $DOCKER_HOST is set correctly.
According to the docs, k8s supports Rocket. Can someone please guide me about how to start a local cluster without a working Docker installation?
Thanks in advance.
You need to set three environment variables before running ./hack/local-up-cluster.h:
$ export CONTAINER_RUNTIME=rkt
$ export RKT_PATH=$PATH_TO_RKT_BINARY
$ export RKT_STAGE1_IMAGE=PATH=$PATH_TO_STAGE1_IMAGE
This is described in the docs for getting started with a local rkt cluster.
Try running export CONTAINER_RUNTIME="rocket" and then re-running the script.