IBM MQ FTE: alternative open source solution requested - file-transfer

We have a need: to provide secure, guaranteed file transfer in our distributed network (Managed File Transfer-class solution, MFT). For that purpose we expect IBM MQ FTE to be ok. But it is quite expensive one.
So, is there any comparable open-source enterprise solution to consider instead IBM MQ FTE?

Related

How to setup Rest proxy on kafka clusture (excluding confluent or any other third party licensed )?

Every search returns the Confluent pages for Kafka REST proxy.
Is there any other way to setup with the vanilla application or any open source ones?
I think the community licensed ones are free to you use. Please check the following page https://docs.confluent.io/platform/current/installation/docker/image-reference.html
Also you can consider this : https://strimzi.io/docs/bridge/latest/#api_reference-bridge if you need only a bridge for integration.

The best way to deploy/redeploy PHP code from github to GCP Compute Engine LAMP Stack [Google Click to Deploy]

overflowers!
Can someone please advice me on the best way to continuously deploy PHP code from github to GCP Compute Engine? Specifically to GCP Marketplace LAMP Stack, which is the Google Click to Deploy VM? Here is the link to the market place
Your advice is greatly appreciated!
Click to Deploy (C2D) is an excellent way to test drive solutions but I'm (admittedly somewhat naive but) skeptical that it's a good approach to combine C2D with customization.
That said, the C2D solutions are published and you could, with some work, customize the solution as the basis for your own solution.
In other words, I'd recommend not combining the C2D as-is but to customize the tools that it uses (!) for your needs.
The README explains how the LAMP VM is built (Cloud Build, packer, chef).
Without wishing to in any way impugn your approach, please consider alternative ways to deploy PHP to Google Cloud Platform. Running Apache and MySQL on a VM may be entirely appropriate for your needs but you will need to maintain the OS, Apache, MySQL etc.
If you're goal is to deploy a PHP (web) app that needs a MySQL-compliant database and you want to be more "cloud native", you could consider using:
App Engine or Cloud Run to host your PHP app (see link)
Cloud SQL for the database (see link)
The above would require more initial work but, if you want more flexibility, resilience and less "chore", I think you'd benefit from the investment.
In addition opening up the app like this would facilitate leveraging Cloud Monitoring, Logging, Debugger etc

OPC UA Global Discovery Server (GDS)

What is the best way to implement Global Discovery Server and Certificate Management? Does open62541 support these services or is there some better library that could be used (not necessarily open source)?
There is a sample code for a GDS from the OPC Foundation here. But it is only community maintained and managing devices via Pull and Push can be quite cumbersome with the GDS Client application. But it was used for a while in IOP Workshops and for V1.04 operation with RSA certs should still work according to standards.
There are some commercial GDS offerings in the pipeline by different vendors, given security is in the play I recommend use rather commercial ones.
For implementing and testing the GDS Pull/Push services for a UA server the sample server should be sufficient.

Why would someone pick apache Kafka instead of Confluent?

I'm about to start deploying to production a couple of Kafka cluster in 2 different DCs. My main use is for replication using MirrorMaker. To continuously stream/replicate ElasticSearch and Postgres between DCs in order to have a (near) real-time backup and failover.
What I can't get my head around is this simplest question: should I use Confluent or apache Kafka?
I can see that Confluent adds many niceties but what I don't get it: why would someone pick plain Apache Kafka then? I've seen this answer and it seems clear: "pick Confluent, has way more stuff".
As answered in linked post, you can add whatever external processes you want to Apache Kafka.
Note: You are not picking either or, you are always picking Apache Kafka. Confluent Platform adds on top of, similar to Cloudera's Data Platform, as an alternative consideration.
If you want to connect Elasticsearch and Postgres (via JDBC), both of those connectors are Open-Source under the Confluent Community License, so that would be one potential reason for not using Confluent products.
Other reason: Do you need the "more stuff"? Are you able to get support from elsewhere? For example, AWS support on MSK or IBM Streams or Azure EventHub are not using Confluent Platform (because it's against the above license)
MirrorMaker and MirrorMaker2 are both under the Apache License, so they have no such usage / redistribution restrictions.
should I use Confluent or apache Kafka?
When deciding on deploying a vanilla Apache or a commercially supported product you should think about the O&M (operation and maintenance) timeline and what you gain and lose. Whatever you choose will be very difficult to replace once in production.
I'll also agree with #OneCricketeer's answer.
Do you need the "more stuff"?
I work as a professional services consultant with some Apache products. My advice is keep your stack (whatever it is) as simple as possible. So if you don't need the additional tools and functionality of Confluent, don't use them. It's how they make the product "sticky" (re: vendor lock-in).
Vanilla Apache Kafka
Pro No vendor lock-in or dependencies
Pro Faster updates and feature development
Con No nice dashboards
Con Harder to secure
Confluent
Pro Commercial support and professional services available
Pro More stable with fast and easy security patches
Pro Nice dashboard and management tools
Pro Easier to properly secure
Con Expensive
Con Expect vendor lock-in and frequent up-sells
My Opinion
If you have money to spare and this will be a critical piece of infrastructure I'd recommend buying through Confluent. If you try to avoid paying for them, you'll have to hire someone (expensive) who knows it anyway and you'll have to deal with the patching nightmare of open source projects.
If this is something you want to kick the tires on, can allow for downtime, or think you'll replace in 2 years, I'd just use the Apache Kafka with one of the open source dashboards.

Differences between connectors Mule ESB and WSO2 ESB

Using Mule ESB I noticed that you can connect (via anypoint connectors) applications, databases, web services etc.
Since I am making a comparison between different ESB (Enterprise Service Bus) I ran into WSO2 ESB and reading the documentation it seems that allows to interact only web services (through SOAP communications).
Someone confirms what I wrote? Or WSO2 ESB is flexible as Mule ESB and I'm wrong (if so what are the differences)?
WSO2 ESB also has the concept of connectors which you can use to connect to external applications, databases, file systems and web services hosted in cloud or in internal networks. Here is a webinar which you can follow to get more information.
http://wso2.com/library/webinars/2014/09/esb-connectors-for-on-premise-and-cloud-integration-solutions/
I can't help as my knowledge of WSO2 is limited. What I could do is to recommend you the book open source ESB in action, although outdated its data, the introduction is amazing and the comparison methodology is also good. You could follow the same approach with the state of the art today.
Most of the connectivity to applications, databases, different protocols are already available with wso2 ESB and wso2 product stack out of the box. However there are some connectors that will support integration with on-premises legacy systems and additional protocols. eg: ejb, is08583, kafka, etc.