Mine is a Windows 10 computer.
I ran the following command to install official kubernetes helm charts repository.
helm repo add stable https://charts.helm.sh/stable
The installation was successful as it says.
I wanted to understand where they are installed.
Searched on my home folder, and also ProgramData but could not find.
Any help would be appreciated.
Watching some videos, just found the answer.
Run the following command.
helm env
And you should see all the relevant locations.
And note, this works on ubuntu as well!
These docs describe the helm default file location per operating system.
According to them, your repository files should be under %TEMP%\helm.
Related
pandoc-crossref must match the pandoc version, and also only the 3.10.0 release works on OSX Big Sur. Thus, it is not possible to get pandoc and pandoc-crossref running in a conda environment from the official channel or from conda-forge.
I could easily download the matching binaries from https://github.com/lierdakil/pandoc-crossref/releases/tag/v0.3.10.0 and copy them e.g. to the binpath:
$ which pandoc-crossref
/usr/local/bin/pandoc-crossref
$ curl -OL https://github.com/lierdakil/pandoc-crossref/releases/download/v0.3.10.0/pandoc-crossref-macOS.tar.xz
$ tar -xzvf pandoc-crossref-macOS.tar.xz
$ mv pandoc-crossref /usr/local/bin/pandoc-crossref
But I think that is not a clean approach, because conda will not know that I updated the version for pandoc-crossref.
What is a clean approach for updating a package managed by conda from a binary available on Github?
Update Feedstock
I updated it on the Conda Forge feedstock, which is what I regard as the "cleanest" solution.
How does one do that? First, OP had posted a comment on the feedstock in the PR that they wanted merged. This was the appropriate first step and hopefully in future cases that should be sufficient to prompt maintainers to act. In this case, it was not sufficient. So, as a follow up, I chatted on the Conda Forge Gitter to point out that the feedstock had gone stale and had non-responding maintainer(s). One of the core Conda Forge members suggested I make a PR bumping the version and adding myself as maintainer, and they merged it for me. In all, this took about 10 mins of work and ~2 hours from start to having an updated package on Anaconda Cloud.
Custom Conda Build
Otherwise, there isn't really a clean solution for non-Python packages outside of building a Conda package. That is, clone the feedstock or write a new recipe, modify it to build from the GitHub reference, then install that build into your environment. It may also be worth uploading to an Anaconda Cloud user account, so there is some non-local reference for it.
Pip Install (Python Packages Only)
In the special case that it is a Python package, one could dump the environment to YAML, edit to install the package through pip, then recreate the environment.
We are used to use this repository to download kubelet and kubeadm packages - https://packages.cloud.google.com/yum/repos/kubernetes-el7-x86_64/Packages
but it's unavailable now, I mean it's empty from RPMs since yesterday.
Can't find those rpms anywhere online and also there is not a status page or update from google that they are taking it down.
The kubernetes documentation still points to this entry and it's empty.
ideas ?
The Kubernetes documentation still points to this entry and it's empty.
Actually this page (The Official Kubernetes Installation Guide) explains that.
ideas ?
The guide suggests adding the repo under the the list of YUM repos. I have been installing k8s components from the following source with no issues.
[kubernetes]
name=Kubernetes
baseurl=https://packages.cloud.google.com/yum/repos/kubernetes-el7-\$basearch
enabled=1
gpgcheck=1
repo_gpgcheck=1
gpgkey=https://packages.cloud.google.com/yum/doc/yum-key.gpg https://packages.cloud.google.com/yum/doc/rpm-package-key.gpg
exclude=kubelet kubeadm kubectl
EOF
They seem to be there just fine for me, https://packages.cloud.google.com/yum/repos/kubernetes-el7-x86_64/Packages/29e7806d1d54cc0eea2963f4ab276778526538816c88c963ece7d1a05fd80792-cri-tools-1.0.0_beta.1-0.x86_64.rpm for example. If you just mean the web server doesn't show the file list for those folders, yes because it's huge.
We have a Spacewalk server we use to distribute updates to our CentOS 6 and 7 client systems. None of the client systems have internet access. Only the spacewalk server has restricted access to complete repo sync updates.
We have an issue when it comes time to push out the centos-release rpms. this rpm installs repo config files like /etc/yum.repo.d/CentOS-Base.repo. Since the client system can't get out to the internet, any subsequent yum command displays an error about not enough mirrors.
Does any one know of a way to have either Spacewalk or yum exclude the /etc/yum.repos.d/ directory on install of the centos-release package.
It has to be an automated or configurable method as I currently run a manual command on each of the affected systems when I know that rpm is being pushed out. (/bin/rm /etc/yum.repo.d/CentOS-*.repo). I considered a cron job, but that just didn't seem like the best option.
Thanks in advance.
You can tell yum not to use a certain repository:
yum install --disablerepo=centos-base
or you can tell him to only use your repository by disabling all others:
yum install --disablerepo=* --enablerepo=myrepo
The solution I have seen in the past is to configure yum to not use the default location, but instead a different one. For example, /etc/yum.internal.repos.d/. Then anything that happens in the default location doesn't matter.
I am using logstash-1.4.2, and follow the tutorial here:
http://logstash.net/docs/1.4.2/contrib-plugins
to install contrib plugins for my stomp.
but when I follow the tutorial in "Automated Installation" part, run:
plugin.sh install contrib
I will have the following:
C:\logstash\logstash-1.4.2\logstash-1.4.2\bin>plugin.sh install contrib
Welcome to Git (version 1.9.0-preview20140217)
Run 'git help git' to display the help index.
Run 'git help <command>' to display help for specific commands.
which does not look that my plugins have been added.
Anyone has a solution for this, Thanks.
The current logstash contrib automated installation only works for Linux/Mac OS, which is why this is failing for you. There are a couple of approaches you can take:
There has been some code checked in for doing this on Windows, just
not yet merged into the production release. Check here:
https://github.com/elasticsearch/logstash/pull/1356
You can just download and extract the files into your logstash
directory ( in your case the one above bin). Discussion of that
approach here:
https://groups.google.com/forum/#!topic/logstash-users/KbLwrgC4qSk
so I recently installed the vagrant-aws plugin with vagrant plugin install vagrant-aws, after quite the adventure with ruby versions and dependencies. Anyway, for our deployment pipeline I need to find a way to spin up vSphere machines with vagrant, and there is a vagrant-vsphere plugin on github, but I have no idea what to do with it. I'm really new to ruby/programming/devOps in general, and it doesn't look like vagrant will allow you to install plugins from local files.
I saw in the vagrant docs you can require a plugin in your vagrant file with require 'my_plugin', and I tried the path to the .gemspec file where my_plugin is, but can't seem to figure out how to work this. If anyone has help or pointers to places on the internet I can look, I would very much appreciate it :) I'll be scouring the web for a long time to find a solution.
You should be able to install it after bulding it from source:
mkdir ~/tmpbuild
cd ~/tmpbuild
git clone git://github.com/nsidc/vagrant-vsphere.git
cd vagrant-vsphere
gem build vSphere.gemspec
vagrant plugin install vagrant-vsphere-0.0.1