Can any body tell me how to create standalone Hornetq cluster setup on Remote machines?
thanks in advance..
Related
I am trying to deploy Apache -Druid in docker container. Image is built successfully. All the services including zookeeper are starting normally when Docker Image of Apache-druid is deployed.
Here is my setup, I am deploying Druid docker image on Docker remote host, it using Docker swarm internally. Ihave configured different container name, hostname for each service of Apache Druid. I have configured external network, I found out internally swarm is initiating those service on different hosts. I have configured "link" as zookeeper for Druid services and vice versa.
But, middle-magaer, co-ordinator and Broker are failing to connect to Zookeeper. Following is the error:
org.apache.zookeeper.ClientCnxn - Opening socket connection to server zookeeper/IP Address:2181. Will not attempt to authenticate using SASL (unknown error) 2020-03-19T22:04:05,673 INFO [main-SendThread(localhost:2181)] org.apache.zookeeper.ClientCnxn - Socket error occurred: zookeeper/IP Address:2181.: Connection refused
So I have different services running on Docker network, on different
nodes(Docker on Linux).
These services are part of Apache Druid like middle manager, broker,router etc. These services are part on one single docker
compose file.
Services start but then not able to connect to
zookeeper which is part of Apache Druid package. Found out from my
infra team that these services are launched on different nodes within
network.
I have used defined external network. Also, I am defining
links . How do I configure services to talk to each other. Here is my
docker compose.
Here is my docker-compose file in comment below
Request inputs.
Thanks and Regards, Shubhada
I have fixed this issue y setting druid host to gateway.docker.internal
I was looking at this url.
It says-"If you already have a way to configure hosting resources, use kubeadm to easily bring up a cluster with a single command per machine."
What do you mean by "If you already have a way to configure hosting resources"?
If I have a few Ubuntu machines within my office LAN can I setup Kubernetes cluster on them using kubeadm?
It just means that you already have a way of installing an OS on these machines, booting them, assigning IPs on your LAN and so. If you can SSH into your nodes to be you are ready!
Follow the guide carefully and you will have a demo cluster in no time.
I have 2 clusters of JBoss servers - one cluster of JBoss 4.2.1 and one cluster of Wildfly 10.
I'd like to configure the two clusters to be able to exchange JMX notifications between each other.
While Wildfly 10 uses JGroups with "RELAY2" support, JBoss 4.2.1 uses JGroups 2.4.1, which does not support "RELAY2".
Is there any way to configure the two different clusters to be able to exchange JMX notifications with each other? How can I configure my Wildfly 10 nodes to join my JB4 cluster?
As per RedHat support, there is unfortunately no way to have the two clusters talk to each other as they are using different clustering techniques, even if both rely on JGroups for the communication mechanism.
I have installed a Service Fabric unsecured development cluster on a shared, on-premises VM with firewall turned off. I can connect to it locally (on same VM) via PowerShell, and deploy locally via Visual Studio. However I am unable to connect or deploy to the cluster from any other box on our network, getting the following error message from PowerShell:
Connect-ServiceFabricCluster : No cluster endpoint is reachable, please check if there is connectivity/firewall/DNS issue.
As I said, the firewall is turned off on the machine hosting the cluster. What am I doing wrong?
OneBox deployment of Service Fabric (installed via the SDK) does not support remote publishing.
Template for configuring a shared dev/test cluster consisting of three nodes can be found here: https://azure.microsoft.com/en-us/documentation/articles/service-fabric-cluster-creation-for-windows-server/#download-the-service-fabric-standalone-package
/Mikkel
While going through Redhat Fuse ESB documentation , I found mention of fabric containers as something different from stand-alone container. Are Fabric containers virtual/logical containers?
Link : https://access.redhat.com/documentation/en-US/Fuse_ESB_Enterprise/7.1/html/Deploying_into_the_Container/files/FESBLocateFabric.html
Fabric containers are real JVMs that are started and controlled by Fabric servers. They are not 'virtual' containers but are real JVM processes.
Standalone containers are single JVMs that monitor their "deploy" folder by default to look for artifacts to deploy. You can start a standalone Fuse server by simply running bin/fuse. This server will not contact any other Fuse servers.
A Fabric is a clustered group of Fuse instances. Because the cluster needs to distribute its artifacts according to some configuration it doesn't look at its deploy folder anymore (it ignores the contents) but uses "profiles" which are stored on the Fabric servers.
If you would create a cluster of 3 hardware servers, you would run 3 fabric servers on them.
On the first server, you start Fuse by running bin/start.
Then run bin/client -r 10 to connect to the server.
You now still have a standalone instance. To turn it into a Fabric server run fabric:create --clean --wait-for-provisioning
On the other two servers, you start Fuse the same way, but instead of running fabric:create you run fabric:join with the relevant arguments to have them connect to the first server.
You'll notice that when you look at the administration console of the first server you'll see the other 2 servers as well, and you will be able to start fabric containers on any one of those 3 servers. You can also attach profiles to those containers.