I have one server.
And when I install ovirt I can open web interface at 10.0.0.10.
But, then I try instal host on 10.0.0.10 after that I can`t open 10.0.0.10
So, how I can instal ovirt-engine and host on one server?
Can you give me link to any tutorials?
If you're going to use a single host for both engine and run vms, the supported and suggested configuration is to use Hosted Engine.
See http://www.ovirt.org/documentation/how-to/hosted-engine/
You should follow the official quickstart guide http://www.ovirt.org/documentation/quickstart/quickstart-guide/
This definitely works for oVirt release 3.6 and Fedora 22 Server.
After you add the host in the webgui, it deploys necessary packages and configurations on the host. One of these steps configures and starts iptables. Maybe you first need to disable iptables and reconfigure it and that's the reason why your ovirt-engine isn't available after your deployment is finished. You can read the details which ports to open for iptables in the Quickstart Guide.
If you use CentOS this post maybe also helpful http://blog.mit.bme.hu/meszaros/en/node/193
oVirt Orb is what you are looking for, if you just want to try it out. "oVirt Orb lets you try out oVirt on your own laptop. You can test it and play with it, all without the need to manually install all the components or use multiple hosts or a storage server."
If you have shared storage available somewhere or don't mind setting it up, Hosted Engine (as Sandro mentioned) will give you an Engine Virtual Machine running on a single Host. (Hosted Engine requires NFS, iSCSI, etc.)
Related
I have a micro service written using Java EE + Openliberty. I have deployed this into kubernetes in my Docker-Desktop. I want to know, is there any way to debug this deployed service? I am using Eclipse. Please let me know how can I debug deployed service. Thank you!
I heartily endorse looking at telepresence. I don't use it for real yet, but I've verified it will be very useful in our enterprise once we get some pieces in place.
However, I don't believe you need that complication for your simpler situation. Telepresence will be very nice for injecting NEW code into a cluster without actually deploying it. In your case, you just want to debug the existing code. This allows for a simpler mechanism.
To remotely debug in Eclipse, you need to connect to a debugger port. To do this with code running in a k8s cluster, you have to expose the port to connect to, and make it available on your desktop.
In your image, you have to ensure that your Java command line has the "-Xdebug" parameters, which specifies the debugger port (5005 by convention). You also have to ensure that your Dockerfile specifies that port to be exposed. When that is running in your cluster, you have to use "kubectl port-forward" on your desktop (with appropriate parameters). This will proxy the debugger port in the cluster to a corresponding port on your desktop. You then connect to that port from Eclipse.
If you want to debug your application, you can do it using either squash or telepresence. Here is great article comparing these two solutions
Article also mentions that squash doesn't support eclipse, but notice that is was written in 2018 and since then squash added support for eclipse IDE.
I wonder if there is a best practice/way how to change the network plugin of a running k8s cluster installed by kubespray.
Concrete scenario: in beginning i installed with flannel, now i realized i need need to move to canal.
Tried to find out if kubespray provides an option to remove only network plugin, but doesn't seem so.
Any suggestions, hints?
I want to install packages all over servers at the same time. Finally I will install a cloud service and distributed filesystem for testing purposes?
I found Webmin has Cluster option, but no clue or clear documentation about how to do. Webmin UI is useful for me, I decided to use it whenever I need inventory of any of servers.
All the modules in the Cluster category make use of the Webmin Servers Index module and RPC to control other systems.
Webmin has great cluster documentation. There is cluster option at navigation menu, just start from looking around and reading manual. It will be clearer for you then.
I am reading the docs about service fabric and was also interested to review how to setup a cluster with multiple VM's, but so far I can only guess based on the devclustersetup.ps1 / it's xml file, but I didn't see any docs on it which explains the various configurations and/or API's.
What I would need is how to set up a simple cluster, how to add/remove nodes, monitoring, setting up resource constraints per node etc so I can setup a sample cluster and test few things.
So far I've done these:
installed VC runtime ( as web pi installer fails without it )
installed service fabric and the SDK ( got the installer out of the web pi installers )
tried to change the sample xml, adding multiple hosts, but then with that I ran into the IPv6 only issue in my setup ( see my other question ), so it didnt work out
Thanks
We are working on NanoServer support for Service Fabric. (I am unable to respond to the comment asking about it because I apparently don't have enough points).
Setting up a multi-machine cluster is not supported at this time so you won't find any documentation explaining how to do it.
There will be a public preview of the service in Azure later this year and the platform will be available as part of Windows Server 2016 for on-premise deployments. As those options become available, there will be plenty of guidance explaining how to setup and manage your cluster.
UPDATE: 2016-03-31
Standalone installation on-prem or in another cloud is now available in public preview for Windows Server 2012 R2 and up.
I read about using haproxy and keepalived to avoid spof for haproxy. Is it possible to do this in ec2?
Say I have 2 instances. Each with haproxy and keepalived installed. The VIP shall be aws elastic IP.
In theory this should be possible, there are blog posts dotted around with instructions on how to do it. However I have been trying to test this over the past few days but not had any luck with it.
Amazon blocks multicast on EC2 (Classic) so the primary communication method for Keepalived will not work. You need to install the latest version (at time of writing this is 1.2.13) which has unicast support built in. This should allow you to bypass the multicast restrictions that Amazon puts in place. I think the version of the keepalived package is 1.2.7 in the repos (yum install keepalived) which does not have the unicast patch included..
You should be able to use wget to download the latest TAR, unpackage it and build it from source (./configure prefix=/, make, make install). Make sure you have the GCC package and openssl-devel package installed before trying to configure as it will fail with errors otherwise.
If I get it working in the meantime I will come back and put a link to my blog with the exact steps needed :)