In Talend (particularly MDM Studio TOS and TOP 4.2.3 on Windows 7), how can I export a job so that it can be imported on a separate machine, into an existing workspace.
To export a Job:
Right click the Job, and select Export Job.
To the right of To archive file, click Browse to choose the destination and name for the Job ZIP archive.
Select the Job Version.
Set Export type to Anonymous Job.
Check all Options.
Click Finish.
A ZIP archive will be created at the selected destination containing the Job.
To import a Job:
Right click Job Designs under the Repository section, and select Import items.
Click Select archive file, Browse, and select the Job ZIP archive.
Optionally, check overwrite existing items if you're importing an existing Job.
Check the Job under the Items List tree.
Click Finish.
The Job should now appear under Job Designs.
Related
I created a deployment group and able to create a agent on my server (linux machine) . Target machine was set. All I need was to create a release pipeline, So I created a release pipeline and select artifacts from build ( I already build that).
For Stage I select empty job and then select “Deployment Group Job” , Add my deployment group that I created and save the changes.
I select another task i.e “Copy Files” that copy files from artifacts to “/var/www/html” on my server(linux).
After this I run the release and deploy it. It succeed but what is does, that it copy zip file to the path I provide on my server.
/var/www/html/11.zip
That is not what I want. I want that it unzip the artifact that was built and deploy code to path I mention so that I can run my application there.
I may be choosing wrong task. But all I need it to run my application from my server. If I need to add some different task , what are those ?
If I need to add some different task , what are those ?
Since you need to deploy Unzipped files to the target path, you can directly use the Extract files task to replace the Copy file task in Deployment Group.
Here is an example:
In the Destination folder field , you could input the target path.
Then this task will unzip the zip file and send the unzipped files to the target path.
i am using azure-pipelines to create a build of my react app and then to deploy it to azure app-service now when i build the app all the required contents go it "/build" folder which then i archive and publish to build artifacts with the corresponding tasks. but the issue is in the archive the structure is "/build/{required files}" but i need the structure as "/{required files}" is there any way to achieve it on azure-pipelines.
i have already looked in azure documentation there they show a variable named "includeRootFolder" but even in the link it does not have a description and in the Archive task i am not able to see any option which will set this variable to false.
There is an option "Prepend root folder name to archive paths" on the task "Archive files" and is enabled per default which causes the behaviour you encounter. Just uncheck this option and this should fix your structure
I have continuous integration setup for the web app project in VSTS. Thing is I need to copy some folder, let's call it "Libs"from NuGet package location to the bin folder of the web app. I employed VSTS "Copy Files" task for it. After couple of failings I've figured out how to properly(at least I think so) setup the task, but, alas, folder is not appearing in artifact zip file and I don't understand why.
"Copy File" task runs successfully. But when I click "Artifacts" button on successful build and look inside "drop" container web app zip file does not contain "Libs" folder under bin directory, as I expected it to be.
What am I doing wrong?
Here is line from my "Copy File" task during the build:
2018-05-22T19:24:37.5090638Z Copying
D:\a\1\s\WebApp\packages\ImportantPackage.1.0.0.0\lib\net461\Libs\SomeLib.dll
to D:\a\1\a\bin\Libs\SomeLib.dll
Here is how I actually setup "Copy File" task:
Here is how Publish task looks like:
There is a typo in your Source Folder: ...\ImortantPackage.1.0.0.0\... is probably supposed to be ...\ImportantPackage.1.0.0.0\....
Try selecting the source folder from the explorer by clicking the three dots next to the folder path and selecting the correct folder.
Task Groups are a great way to build a library of commonly executed Build and/or Deployment Actions in Visual Studio Online. Does anyone know of a way to make a copy of (or clone) an existing Task group? I don't see that as an option in the UI anywhere.
I think there is no way to do this yet. It's not even possible on TFS On-premise. Though our TFS professional told us, it should come with TFS2017 Update 1.
Only thing you can do, is select all tasks in your task group and create a new TaskGroup with them. Which is basically the same as cloning the task group itself - it only takes you some clicks more.
Or you could export your task group, change its name and Id in the exported Json, and import it.
Currently there isn't the copy Task Group feature in VSTS and TFS.
You can create TaskGroups according to your requirement and manage them from TaskGroups tab, then use them directly in different scenario. Please follow below steps.
Create a Task Group from Build/Release definition
Navigate to Build&Release > Task Groups
Select the Task Group you need to modify, change the Properties/Tasks
Save the Task Group
If you just want to copy the Task Groups, you can install the "Copy Build Steps" extension. it can copy build/release steps from one definition to another, also support copy the Task Group.
After installing the extension, navigate to the Builds hub > All definitions. A new menu item, ‘Copy build steps’ will be available when you click the ‘’…’’ behind the definition.
To Use a task group:
In a build/release definition, choose Add steps/tasks to open in the Task catalog. Select the category you specified for your task group and find the task group (it displays the name and description you specified when you created it), then choose Add - just as you would with any other task.
We're using an on-prem VCS and CI pipeline, and don't have plans to switch to VSTS right now. However, I'd be very interested in running cloud-based load-tests against our app as part of our CI pipeline. In order to do this, I'd have to be able to programmatically upload the loadtest script and invoke it from VSTS.
Is this possible?
Yes, it is possible, the workflow like this:
1.Create a valid loadtest file. You can use the load test file from an earlier run through Visual Studio for this.
2.Create a location to upload the file(s). This location is a drop folder on Azure Blob and is below referred to as "TestDrop".
3.Upload the loadtest file and any other files required for the run, this includes the webtest files, settings file, etc. to this location or "TestDrop".
4.Create a Test Run using the Testdrop from the previous step as all the files required for a run are now available at the drop location.
5.Start the run.
6.Once finished, download the results to your local machine. This will be a gzip file. Uncompress it to get the results file.
7.Use Visual Studio to view the downloaded results.
More information, you can refer to this article, which contains samples.