Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 2 years ago.
Improve this question
how to integrate jmeter script with azure devops pipeline?
I am new in azure devops,give some insight view for this?
I have pushed the jmeter code to azure devops repo and then created build for that but afterthat how can I integrate with pipeline.
Not very sure which exactly part you want to know, here just provide you the basic application in VSTS.
Until now, there still has 3 load testing tasks can for you use to integrate the Jmeter: Cloud-based Load Test, Cloud-based Web Performance Test and Apache JMeter Test. These 3 tasks can all be used in CI and CD.
But, there one thing you need pay attention that we are gradually deprecating Cloud load testing(CLT):
Existing customers of CLT:
For existing Azure DevOps customers using cloud-load testing
functionality in the Azure DevOps portal, this feature will no longer
be available as of March 31, 2020.
For existing Azure customers using the performance test feature in App
Services or Application Insights, this feature will no longer be
available as of March 31, 2020.
For Visual Studio users leveraging cloud-based load testing, this
feature will no longer be supported as of March 31, 2020. Visual
Studio Enterprise customers can, however, continue to run load tests
on-premises using Test Controller/Test Agent and will be supported for
any issues that may arise during the support lifecycle of the Visual
Studio version.
New customers of CLT:
New Azure DevOps organizations and new Azure subscriptions created
after March 31st, 2019 will not have access to the cloud-based load
testing functionality.
Since until now, we are still support these tasks, you could see and choose them in Deprecated tasks panel:
Then you can follow above task document to configure your test task.
Below are another blog and docs you could take as refer:
Configuring Jmeter Tests in VSTS/TFS (Azure DevOps) and publishing
Results
Run Apache JMeter load tests with Test plan
Using a self-service model for load test and cloud-load testing,
e.g. Apache JMeter
The pattern described in the post can be considered as well: https://medium.com/#maninder.bindra/load-testing-in-azdo-pipeline-using-dynamically-created-azure-container-instance-with-apache-bench-d4c3e1f9dc
In this approach a load inducing Azure Container instance is created (as part of pipeline, with Apache Bench or JMeter installed, and in a different AzDO agent pool) for the duration of the Load Test, and after the Load test results have been recorded the ACI instance is deleted.
It has been indicated in the post where you would need to make changes to the pipeline yaml if JMeter needs to be used instead of Apache Bench. Hope it helps.
Related
The goal I'm trying to accomplish is migrating out of CVS to using Azure Repos. Currently BAR deployments are sent through CVS repos and then deployed to IBM Integration Bus. I would like to accomplish this same process through Azure DevOps.
I know this extension exists:
https://marketplace.visualstudio.com/items?itemName=ms-vsts.ibm-integration-bus&ssr=false#overview
However there is limited documentation available. I'm curious to know if anyone has had success using the above extension?
If not through the extension is there another solution available?
I have never used that extension myself.
The tasks involved in setting up a build and deploy pipeline are not difficult, so you should be able to roll your own.
See mqsicreatebar and mqsipackagebar for how to build the BAR files.
See mqsideploy for details of how to deploy the BAR files.
The rest will depend very much on your chosen source code control system and build orchestration technology.
Forgive me for asking a stupid question. I am from IT Infrastructure background & have been asked to create CI/CD pipelines based on my recent learnings on DevOps.
We have couple of applications whose source code is currently in TFS 2013 & those apps are written in ASP.NET C# language. Now, requirement is to migrate the source code from TFS to Azure Repos (Azure DevOps services) & further create a CI/CD pipeline.
Now for demo purposes, customer is asking us to do the deployment (i.e. Release pipeline) on a test server which is a plain windows 2012 OS without any SQL & IIS for both of these applications. Is that possible & how could we achieve the results to confirm release pipeline is funcioning properly?
In my opinion, it wont work as there is no application infra/configuration done for those applications on that plain test server. I guess we actually need a ready dev/stage environment which is replica of production to do the testing of release pipeline for those applications. Am I correct?
Just need expert advise for confirmation so I communicate the same to customer.
Azure DevOps Pipelines use an agent to perform the deployments. You can run the agent entirely in the cloud when deploying to Azure resources. You can also install an agent locally. Follow this link and scroll down to read about self-hosted agents. This is how you can deploy to your test instance from the pipeline.
Now, what you deploy there may require additional software be installed. You say it's an application in C#. Cool. Now, what's it do? Is it a windows program? Then just having the server there, with an agent installed, is all you need. Is it a web program? Then, yeah, it's going to need an IIS (or whatever) instance available somewhere to deploy to. Is it a database program? Then, yeah, it's going to need a database instance to deploy to. There's nothing magical about having a VM or a machine somewhere. All the same rules have to apply. There has to be an OS, drives, memory, and yes, supporting services depending on the needs of the application.
However, using a local machine instead of a hosted one, that works fine. Just follow the instructions in the link above.
I am doing some research for Powerapps integration with Azure DevOps.
However there is limitated information for it.
It is possible to integrate powerapps inside a Task for AzureDevops?
Based on, that we have a .zip file with the Powerapp, and we want to create a Build and Release/Deploy for several environments.
Thank You.
It is possible to integrate powerapps inside a Task for AzureDevops?
Yes it is.
You can leverage the Solution concept of the Microsoft Power Platform and the Power Apps BuildTools (preview) extension for Azure DevOps.
Update 11/2020: This is now GA and called Power Platform Build Tools
I've written a complete step-by-step guide on this topic:
A Continuous Delivery Approach for No-Code Solutions in Microsoft’s Power Platform
Bottom line:
With this build tool, you can automatically check-in a Solution into source control and deploy it using a continuous delivery approach with the help of Azure DevOps. See the screenshot for a sample configuration of the Export and Import Solution Task.
It works for everything you can organize inside a Solution, e.g.:
Power Apps
Power Automate Flows
AI Builder Models
Common Data Service Entities
It is possible to integrate powerapps inside a Task for AzureDevops?
I am afraid there is no such Task integrate powerapps for AzureDevops at this moment.
If you want to integrate powerapps with azure devops, you can follow the guide step by step:
Microsoft Teams – Integration with Visual Studio Team Services using PowerApps.
Besides, AFAIK, PowerApps should not be "Build/Deployed" through Azure Devops.
When you are developing with PowerApps, there is no way to do Source
Control. There are no source files. The only artifact you can version
control is the .zip file that you can export.
And
In PowerApps, you don’t have to build your code. Any change you make
to the application is live for you to test it. In that way it is very
productive. To publish the application you just click on the publish
button and it is live.
Check this great blog: PowerApps From A DevOps Perspective for some more details.
Hope this helps.
Solutions are a way to package your components in a single zip file and use Powerapps build tools to import your solution on to a different environment or tenant.
It is still a an improvement from manually importing each app or environment variable and then import it on to target system, but it lacks what we call as automation of deployment.
To provide an eg, I will explain what I have done, and what still constitutes of a manual task:
I created an enterprise level app using powerapp canvas model. My app consumes data from around 20 APIs. These API calls are implemented in power automate.
We have 4 environments, dev, sit, uat and prod. Now I cant keep on importing flows in each environment and change their api URLs to point to the deployed environment. So I used environment variables for each environment which stores api URLs for each environment. This can be done under solution.
Under the same solution, I added my app. So now my solution has 2 things, my app and the environment variable which consists of api URLs.
I then use powerapps build tools to move this solution from dev to sit.
Steps: use build tools tasks to perform the following
Export solution
Unpack it in git
Pack it
Import the solution.
This successfully moves my solution to sit.
But the solution environment variable still points to the dev url.
So I have to override environment variables to store sit URLs.
This manual intervention to edit environment variable is as good as doing all the tasks manually.
This was the case when PowerApps was first announced; however, this is no longer the case.
While it is technically true that there is no actual code that would be managed and deployed with a PowerApp or Flow but that doesn't mean that you can not use the power of Azure DevOps. Additionally, when creating a PowerApp / flow you would also be creating entities and even Model Driven apps - and these uses solutions - which naturally work well to deplooy within Azure DevOps.
Microsoft is building out this whole construct to enable all these to deploy...
While the whole incorporation of PowerApps and flows into Solutions is not fully baked yet - they are targeting to have this ready around the October time frame this year.
We have been talking to Microsoft about also enabling PowerApps and flows to follow the same expansion that solutions do so that they can take advantage of the full branching strategy.
So even though you would be simply exporting out zip files into your repo - you can still take advantage of the full devops pipeline which is highly recommended.
Use this component, it still on preview mode but is working fine on my side
https://marketplace.visualstudio.com/items?itemName=microsoft-IsvExpTools.PowerApps-BuildTools
I have a Visual Studio (formally MSDN) Enterprise License.
My newly created Azure DevOps Service does not reflect this when looking at the self-hosted pipeline configuration under Project Settings > Pipelines > Retention and parallel jobs.
According the the (i) info button:
Visual Studio Enterprise subscribers get one self-hosted parallel job
as a subscriber benefit.
Anyone know how I can get my Enterprise Subscription to show through to Azure DevOps Services (formally VSTS)? It was pretty simple to do with TFS, but am drawing blanks on ADOS.
My user account for both my Enterprise account and ADOS are one and the same - I was rather hoping it would just show through - apparently that's not the case.
The best I found was this page - but it's only for TFS and explicitly warns that:
The requested page is not available for Azure DevOps Services. You have been redirected to the newest product version this page is available for.
My initial attempt at asking MS through their online chat "Concierge Service" was met with apathy and a suggestion I ask elsewhere...
Apparently patience is a virtue.
It took several days but eventually (or was it because I also raised this issue on the Azure Developer Community site?) it correctly shows my associated Enterprise Subscription.
We have Continuous Integration setup through TFS 2012.
Requirement:
Wanted the release build to be deployed on mulitple servers as a part of the CI. We need to do a xCopy of the files onto the server folders.
On TFS with Azure there seems to be a facility to publish it directly onto the server (http://www.windowsazure.com/en-us/develop/net/common-tasks/publishing-with-tfs/)
There-in it uses 'AzureContinuousDeployment.11.xaml' as the Build Process Template.
Wanted to check if anybody has tried similar publishing for custom/local servers, maybe using Custom deployment template, etc.
You likely want to create Standard Environments for Automated Deployment and Testing which you can do out of the box with TFS 2010. TFS 2012 and TFS 2013.
You use the "Lab Centre" which is part of the Microsoft Test Manager product to configure Environments that consist of one or more target machines.
You then use the Lab Default Template to push a build output to the target servers and execute a script to complete the install. You can even have your Integration and UI tests executed as part of the same actions.
If you are looking for more integrated Release Management with Team Foundation Server then you should look to Octopus Deploy as a good way to go all the way to production.
We use it with many of our customers and it works a treat...
From my understanding TFS 2012 ships with a Lab Deployment template. You configure the lab definitions of which servers are involved and then configure the deployments to those servers. Also used for hooking Test Manager in to run unit tests in the lab.