I am trying to figure out how to import my external test results into Azure DevOps.
This article describes how to publish the results within the same pipeline. But that's not going to help me. We do use Azure pipelines to build and deploy our solution, but the testing is run after that (in a deployed environment), so, outside of the pipeline. We can collect the results (in a Azure DevOps suitable format) and would like to feed it back into Azure DevOps release that did the deployment.
All tips are welcome.
Related
I have deployed the NetCore application into Azure through the Azure DevOps CI/CD pipeline. In the same code pipeline, I want a PDF to be automatically generated and output to a publicly accessible directory.
And specific sections of the documentation are converted into markdown and placed in the appropriate developer portal folders.
I am having all my codes in Azure Databricks notebook. I want to automate execution of my notebook using azure devops.
I have linked my Databricks Repos with Azure Repos. Created a build pipeline to create artifact of my code. In release pipeline.
I was able to deploy my code to lower environment(Here same Databricks workspace where I have my code).
But I was not able to execute/Run notebook directly from Azure Devops.
How do I deploy code from my lower environment to higher environment(Prod Server) Databricks workspace.
To deploy code to higher environment Is it same as deploying to lower environment. Just changing details of workspace.
Like Databricks Token, Workspace URL, etc.
You need to create release pipeline and push your code to GIT repo. Follow this Microsoft documentation for details -
https://learn.microsoft.com/en-us/azure/databricks/dev-tools/ci-cd/ci-cd-azure-devops#:~:text=The%20release%20pipeline%20deploys%20the,Pipelines%20menu%20and%20click%20Releases.
This demo shows how to do that. The steps include:
Testing
Update checkout at Databricks workspace - for example, you may have a separate checkout for testing. You can do that via Repos REST API, or via databricks-cli (databricks repos update command)
Triggering execution of tests by using the Nutter library
Reporting testing results to DevOps
Release
If tests are successful, you can push updates to staging/production environment - this could be done via DevOps Release pipelines
You can use the same Repos REST API or databricks-cli to update checkout in staging/production
I am looking for the best way to create CI/CD pipeline for GCP BigQuery using Azure DevOps.
I need repositories in Azure Devops, and automatically have CI and CD to our Dev/QA and PROD environments in GCP. Also, how to set up automatic builds in Azure DevOps for GCP master branch code? Is there a way to integrate BigQuery and Azure Devops repositories ?
We have to use Azure DevOps for Project, repository and pipeline management. So trying ways to build and deploy to GCP Bigquery from Azure Devops. Any insights would be helpful.
First of all congratulations on your first post.
Yes definitely you can integrate Azure Devops with BigQuery.
A good read here.
But from your question, I failed to understand what you are trying to deploy. Is it some client code which will connect to BQ to run query and fetch data or is it a Dataflow job starting from ingesting data?
There are different ways to do each of them.
I have an ASP web application that I have building in an Azure DevOps Build Pipeline. That is all fine.
I want an external organisation to be able to define their own Azure DevOps Release Pipeline to consume the build artifacts produced by our Build Pipeline. I need the access of that external organisation to be restricted with some sort of credentials (i.e. I don't want the project to be public to everyone). The external organisation should be able to deploy the latest version.
I thought this would be a relatively simple process using only Azure tools (particularly with reference to Feeds), but have tried a number of different approaches based on the documentation but all have failed. I don't want to publish to GitHub - I just want to keep everything inside Azure. I have tried using Universal Packages with Feeds, but the Release Pipeline can only pull a specific version from the feed rather than LATEST.
Does anyone have any recommended approaches I should take?
There is Latest option in the feed of your Release Pipeline:
I'm migrating Teamcity to Azure Devops, as with our on-prem infrastructure. During that period we'll have both Teamcity and Azure Devops deployment for different infrastructures, but we want to reduce the amount of human intervention. So I was thinking in having either Teamcity triggering the Azure Devops build, or vice Versa. I think it would be easier to implement a powershell script that triggers the build on Azure Devops, instead the other way around because of compatibility issues, or version issues(old version).
Yes, you can use Azure Devops api to trigger the build. You would use this api call with appropriate body:
POST https://dev.azure.com/{organization}/{project}/_apis/build/builds?api-version=5.0
Reading: https://learn.microsoft.com/en-us/rest/api/azure/devops/build/builds/queue?view=azure-devops-rest-5.0
I'm not familiar with Azure Devops powershell modules if any, but they might work as well.