re-IP'd Chef Server and now can't authenticate - certificate

We've recently moved our Chef server from 'dev' to 'prod' and basic things (knife node list, for example) seemed to work OK.
This morning I went to bootstrap a new node and received the following error:
$ knife bootstrap <IP> -N devops
...
================================================================================
Chef encountered an error attempting to load the node data for "devops"
================================================================================
Authentication Error:
---------------------
Failed to authenticate to the chef server (http 401).
Server Response:
----------------
Failed to authenticate as 'devops'. Ensure that your node_name and client key are correct.
...
Do I have to somehow regenerate keys on the server to complete the move to 'prod'?

Related

GitHub Actions - Upload Artifact Failed: Unable to get local issuer certificate

We are using self-hosted runners (Windows) for GitHub actions. Recently, our company changed the proxy. We have updated the System Environment Variables to the new proxy, and we have updated the proxy details in .env file as suggested by GitHub documentation. And we have added the Root CA in the Windows Certificate Manager (Certificates - Local Computer -> Trusted Root Certification Authorities -> Certificates). But when uploading the artifact during workflow run, we got the below error message and workflow is failing.
Create Artifact Container - Attempt 1 of 5 failed with error: unable to get local issuer certificate
Error: Create Artifact Container failed: unable to get local issuer certificate
##[debug]Node Action run completed with exit code 1
How can we resolve this error? Is there anything that should be changed in runner configuration?

Azure DevOps on-premise cannot verify Kubernetess service connection

I am creating New Kubernetes service connection in Azure DevOps Server 2020 Update 1 via KubeConfig.
When I click to Verify that the connection it says that Verification Failed with the generic error:
Failed to query service connection API: 'https://ekm.mpu.cz/k8s/clusters/c-qmcrb/api/v1/nodes'. Error Message: 'An error occurred while sending the request.'
Please note that the Kubernetess instance is in the other domain.
I have the notion that the error could be with the certs are not imported somewhere on the machine, where the Azure DevOps is hosted, but I am unsure where. The MS documentation is silent about that as well.
So far I've tried to:
Import CA certs to the MMC under trusted publishers.
Import CA certs under cacerts in JAVA-HOME via keytool.
Import CA certs into azureTrustsStore.jks in JAVA-HOME via keytool.
For all 3 I've checked that the CA certs are imported correctly. But to no avail. Could you please advice or redirect me to the method, how to do it?
Additional Info:
While I cannot Verify and Save the connection, I still can Save it and then use it in the pipeline and it works OK! (sucesfully connect and execute the command).
Connection issues can occur for many reasons, but the root cause is often related to an error with one of these items: Network, Authentication, Authorization. You may refer to Basic troubleshooting of cluster connection issues for detailed troubleshooting steps.

access denied: channel [] creator org [Org1MSP] in hyperledger fabric with kubenetes?

I have created a fabric network with kubenetes. All my peers are up and running.
When go to cli shell and execute below command
peer channel list
I get error as
access denied: channel [] creator org [Org1MSP]
On peer logs i get below error
channel []: MSP error: the supplied identity is not valid: x509: certificate signed by unknown authority
I thought its a certificates issue. So i tested same certificates with docker. but it seems that on docker its working. so let me know what is the issue here ?
CORE_PEER_MSPCONFIGPATH
was missing in your peer's environment variables.
Please set CORE_PEER_MSPCONFIGPATH=<msp-dir-of-peer>

TFS Build Agent Pool: Error Connecting The Server

TFS Build Agent Pool Setup Error:
I am trying to setup an agent pool on a server on a domain to connect to the TFS which in another domain,
Expected:
A prompt appear to enter the Authentication for the user name & Password for the TFS
What Happen: Error:
Error Connecting The Server, An error occurred while sending the request.
The PowerShell Script:
PS C:\agent> .\ConfigureAgent
An existing configuration file was detected. This will update the local agent settings. Do you want to also replace th
e server registration (default is N)? n
Enter the name for this agent (default is Agent-UKWTSVULM768) -------
Enter the URL for the Team Foundation Server (default is ) https://tfspro.-----.-------.net/tfs
Configure this agent against which agent pool? (default pool name is 'default')
Enter the path of the work folder for this agent (default is 'C:\agent_work')
Would you like to install the agent as a Windows Service (Y/N) (default is N) n
ErrorConnectingToTheServer
An error occurred while sending the request.
PS C:\agent>
I just tested on my side to deploy a build agent cross domain. After enter the TFS domain username and password, I got a TF14045 error.
After seeing this error, run .\RunAgent.cmd, you'll see the agent register successfully.
Another way you can try is to create a shadow account like case How to authenticate build.vnext agent to on-premise TFS do.

Handshaking with Octopus server fails from a TeamCity agent

We've been successfully using TeamCity and Octopus on a single machine, and now we have added a new build agent due to growing number of simultaneous builds. I have set up a new Octopus Tentacle and verified that Tentacle is reachable from the Octopus server in Listener role. However,any build that uses Octopack fails when executed on the new TeamCity agent. Here is an extract from the build log:
Running command: octo.exe create-release --server
servername:8443 --apikey SECRET --project NRK Bridge (Legacy)
--enableservicemessages --version 0.0.0.662 --deployto Dev --waitfordeployment --package=NRK.Bridge.Web:0.0.0.662 --packageversion=0.0.0.662
[12:06:02]Creating Octopus Deploy release
[12:06:02]Octopus Deploy Command Line Tool, version 2.0.8.22
[12:06:03]Handshaking with Octopus server: servername:8443
[12:06:06]The following certificate errors were
encountered when establishing the HTTPS connection to the server:
RemoteCertificateChainErrors
[12:06:06]Certificate subject name: CN=servername
[12:06:06]Certificate thumbprint: 79C26(...)
[12:06:07]The following certificate errors were encountered when
establishing the HTTPS connection to the server:
RemoteCertificateChainErrors
[12:06:07]Certificate subject name: CN=servername
[12:06:07]Certificate thumbprint: 79C26(...)
[12:06:08]The following certificate errors were encountered when
establishing the HTTPS connection to the server:
RemoteCertificateChainErrors
What I noticed is that the certificate thumbprint in the build log is different from thumbprints shown in Octopus server machine page. I tried to generate a new API key and re-register a Tentacle. Am I overlooking something here?
Did you copy the Api key into Team City? The api key is different to thumb prints. Api Key can be found in Octopus Admin section, under users tab.