I have an Hybris repo on Azure Devops and wanted to deploy on SAP Commerce Cloud(CCv2).'¨
I'm new on Azure Devops so not sure how to proceed, how to connect from Azure to CCv2 etc.
I have already created "Staging" and "Prod" spaces on CCv2 and now I want to deploy only on staging.
If someone already created this kind of task before, is it possible to explain me the steps one-by-one.
For example how can I connect from Azure to CCv2? With some commands or security files etc?
Thanks!
Related
we are using Azure Pipelines to build and deploy various projects. We are using Enterprise Github service connection as our git repository source which we are using as part of the build pipelines that are running on on-prem AzureDevops build agent. We are then using on-prem hosted DevOps agents to deploy the service on our on-prem servers.
If we trigger a build we then receive an email notification about all the new commits which are included in the build. Is the same information (what commits are part of the build) included somewhere in the Azure DevOps UI? I was not able to find it and it's really difficult to then determine what I'm actually deploying our various environments as I'm not always the one who triggers each build.
Thanks
I am looking for the best way to create CI/CD pipeline for GCP BigQuery using Azure DevOps.
I need repositories in Azure Devops, and automatically have CI and CD to our Dev/QA and PROD environments in GCP. Also, how to set up automatic builds in Azure DevOps for GCP master branch code? Is there a way to integrate BigQuery and Azure Devops repositories ?
We have to use Azure DevOps for Project, repository and pipeline management. So trying ways to build and deploy to GCP Bigquery from Azure Devops. Any insights would be helpful.
First of all congratulations on your first post.
Yes definitely you can integrate Azure Devops with BigQuery.
A good read here.
But from your question, I failed to understand what you are trying to deploy. Is it some client code which will connect to BQ to run query and fetch data or is it a Dataflow job starting from ingesting data?
There are different ways to do each of them.
At the company, we will change the URL for the Azure Devops workspace from https://oldname.visualstudio.com to https://dev.azure.com/newname. We have a lot of connections to Slack in Azure DevOps using two approaches. Older links via service hooks in Azure Devops, new links are via slack applications Azure Repos and Azure Pipelines.
Service hooks in Azure DevOps use Slack's Incoming webhooks, so they will not be affected. However, Microsoft prefers to use its Slack Azure Repos and Azure Pipelines extensions, and the way they work is not clear. E.g. Azure Repos is set up in Slack via /azrepos subscribe [https://oldname.visualstudio.com/projecturl] and this step automatically set the address to azdevchatops.azure.com in Azure Devops Service hooks settings.
So it looks like you need to reconfigure all subscriptions in Azure Repos and Azure Pipelines after changing the Azure DevOps URL. But that is a presumption. It is not mentioned in the documentation and that is why I ask. We have enough connections and that means a lot of manual work to configure everything in Slack via /azrepos subscriptions and /azpipelines subscriptions again.
Does anyone have the experience that when using Azure Repos and Azure Pipelines as Slack Apps, everything has to be reconfigured in Slack as soon as the Azure DevOps workspace address changes?
Notification integration is seamless. Notifications work when you change the name (and URL) of Azure DevOps organisation.
I tested just now and everything is ok.
Note: there are another problem. After Azure Devops organisation URL change you need to reinstall/register build agents again to the new URL. If you use hosted Azure Pipelines in cloud, you need in Project Settings > Agent Pools > Azure Pipelines > Agents disable Hosted Agent and enabled again here. Otherwise, CI / CD pipelines do not work.
Our devops team have created an application (ex:athena) and registered with AD. They also have given us Service Principle.
The question I have is , how do I deploy my code in GitHub to the application (athena) that the devops team created for me using the “Deploy to kubernetes service” configuration template ?
Apologies in advance, as I am not proficient in Azure and this shows my gap in understanding.
The Information I have are :
Repository - GitHub (I have done the appropriate authorisation and can see the repository)
Service Principle (create by devops team)
Application (created by devops team)
I have created a Container Registry and Kubernetes service using azure portal
Now, I want to use the “Deploy to kubernetes service” configuration template.
Help much appreciated.
If you want to use this Deploy to kubernetes service, you must get two service connection : Azure Resource Manager and Kubernetes Service Connection.
So, first, you need to configure the connection between Azure Kubernetes, ARM and Azure Devops. Enter project setting->service connection, open New Service Connection and select Kubernetes. Input the relevant configure according to your Azure:
And so do with Azure Resource Manager, you can follow this doc to configure it. Here is my ARM connection you can refer:
Then, you can begin your build and release pipeline.
Deploy to kubernetes service task used in release pipeline. In build pipeline, you must run docker build, push task to finish pushing to Azure registry Container.
And then, run this deploy task in release. You can refer to this blog which written by Azure DevOps Labs: Deploying a multi-container application to Azure Kubernetes Services. It has detailed steps you can refer.
In addition, there has two build source type. One is you import your github repos into Azure Devops repos. So the build can be trigger by Azure Devops Repos. And the other is select Github as your build source, in this type, you can triggered directly by your github instead of use Azure Devops repos:
And also, this need you get the service connection with your Github first. Then, authorize it during the build pipeline.
How to migrate on premises Jenkins jobs to Azure DevOps? Is any plugins available?
I know manually how we can create Azure DevOps pipeline.
There is some available guidance from Microsoft on migrating from Jenkins to YAML builds in Azure Pipelines, but there are no plug-ins or tools that will automate this for you at the moment.
There is, however, a way to trigger Jenkins builds from Azure Pipelines using the Jenkins Integration extension. This may be a good option as you transition.
No, there are no plugins available for that.