Generating Jenkinsfile for (Spring) Cloud Pipeline - spring-cloud

I came across Spring Cloud Pipeline (which is now Cloud Pipeline). The docs are not very clear and I have gone through the new ones for scripts and jenkins projects.
What I would like to do is generate the opinionated Jenkinsfile for my existing repo (Spring boot/gradle project) and deploy to my existing Jenkins. How can I do that?
Thanks

Related

Is there a way we can deploy all the cloudformation templates from a gitlab repository using gitlab pipeline in aws in a single stack?

I'm looking for an option to pick all the templates from the repository without hardcode the yml template files and in future if new templates are added, the pipeline should automatically pick all of them and do the deploy and create a single stack in aws environment, without making any modification to gitlab-ci.yml/pipeline file.
I tried using deploy CLI command, it deploy all the templates but then it goes for update and start deleting one by one and only the last template resource will be available after the pipeline execution is complete.
Let me know if there is an option to do this?

Best practice for automating cloudformation stack in github actions when repository holds cloud formation templates of multiple services

We have a repository in GitLab which stores the cloud formation templates of IAM roles, permissions, security groups and Route 53 of multiple micro-services. Previously we used a separate repositories to handle the deployment from the repo above using Circle CI and Python scripts, to write ONLY the modified artifacts to s3 and then trigger Cloud formation stack to do the deployment after confirmation. for a visual representation, it looks something like this.
I am currently migrating this to GitHub and moving away from Circle CI, and would like to update the CI/CD here to follow best practices and to use GitHub Actions to do the deployment of the infra. What would be the best means to do so?

AWS Glue - version control and setting up for continuous integration

We are in the process of setting up the CI / CD process for AWS Glue ETL Process. The existing ETL process contains the following AWS Glue Components - Crawlers, Registered tables in catalog, Jobs, Triggers and workflows.
Obviously the first step is to set up a code repository and link the existing artifacts from different components mentioned above to the repository, which will ideally need to facilitate the developers in performing the check-ins and pull request from the tool (Something similar to ADF and Databricks). However as far as we have explored, AWS glue does not have integration to any of the source code repository which can directly provide this feature unless we are missing something.
Hence what is the method to setup the environment for CI (I'm still not talking about CD), the below link gives a reference for CI/CD:
https://aws.amazon.com/blogs/big-data/implement-continuous-integration-and-delivery-of-serverless-aws-glue-etl-applications-using-aws-developer-tools/
However it mentions at the beginning that, AWS CloudFormation template file for deploying the ETL jobs are both committed to version control - so not clear on how this is done for the on-going regular commits from the developers.
However as far as we have explored, AWS glue does not have integration
to any of the source code repository which can directly provide this
feature unless we are missing something.
Correct, Glue does not have VC integration.
I develop (python and cloudformation) locally on vscode and use it's git integration plugin. And I use a container if I want to test something locally, but Glue also has a Dev Endpoint for similar tasks.

Terraform CICD pipeline GCP

I am new to terraform and I want set up a CI/CD pipeline to GCP with github to replace a current system that use's jenkins, as we want to increase automation of deployments. What would be the best way or architecture to do this.
One of the primary products related to CI/CD is Google's Cloud Build.
https://cloud.google.com/build
It's one liner reads:
Build, test, and deploy on our serverless CI/CD platform.
It has built in triggers that include GitHub integration meaning that when events occur on GitHub, Cloud Build runs its prescribed recipes.
I'd suggest reading the documenation found at the above page and also correlate against the curated documentation found on GCP Weekly here:
Tag: CI
Tag: Cloud Build

How to Azure DevOps CI/CD Pipeline for PowerBuilder 2017 R3 project, is it even possible?

Summary
Recently migrated PB126 apps to PB2017 and changed source control to Azure DevOps Git.
Now, I'd like to integrate Azure DevOps CI/CD Build Pipeline to the app dev life-cycle.
jenkins
I know it's feasible to configure jenkins CI server so it builds PB2017 projects.
Continuous Integration with PowerBuilder 2017, Bonobo Git and Jenkins
My problem here's I can't get it to work on a local Docker container and make it accessible to the outside world (Internet) so Azure DevOps can trigger its build action. Supposedly, it's a Docker for Windows thing which Docker handles differently from the Linux-based Docker.
Azure DevOps Pipeline
As per this link, Azure Pipelines is the CI/CD solution for any language, any platform, any cloud, it says any language, which makes me believe it's feasible to build PB2017 projects using Azure DevOps Pipelines.
The fact is that I'm totally new to CI/CD in terms of implementing it myself. I've experienced it in many projects where I wasn't responsible to implement it. Now I am. I've been on it since a few days now, and I do want it to work.
Any help appreciated.
The Appeon offical user forum: https://community.appeon.com/index.php/qna/q-a