I have Vagrant 1.6.3 and VirtualBox 4.3.12 on Windows 8.1 Pro(64 bits). Vagrant was working fine till yesterday.
Suddenly, it started giving this error on vagrant up (in powershell):
The provider 'virtualbox' that was requested to back the machine
'default' is reporting that it isn't usable on this system. The
reason is shown below:
Vagrant has detected that you have a version of VirtualBox installed
that is not supported. Please install one of the supported versions
listed below to use Vagrant:
4.0, 4.1, 4.2, 4.3
I have tried following:
Uninstall both Vagrant and VirtualBox (5-6 times) and reinstall. I rebooted after every install/uninstall.
Tried different (latest as well as older) versions of both Vagrant and VirtualBox. I have tried Vagrant 1.6.3 and Vagrant 1.6.5, VirtualBox 4.3.16, 4.3.12, 4.3.10.
Disabling anti virus and firewall. (Both Vagrant and VirtualBox are added to exclusion list)
Checking that VirtualBox is added in Environment Path.
Anything that I am missing? I tried searching and only two relevant links I found were: Vagrant has detected that you have a version of VirtualBox installed that is not supported (does not work), Vagrant has detected that the VirtualBox installed is not supported (fixed in Vagrant 1.5)
Any should I do now?
The problem was with Firewall. I recently switched to Comodo firewall and it was causing the issue.
Related
I have a google VM instance which was running flawlessly. It has Centos-7 and Plesk installed in it.
I just stopped it, upgraded machine type (to better CPU and RAM) and started it again. My server stopped responding at all. No websites are running, I can't connect to SSH & Google Cloud SDK Shell is unable to reach server. It says NETWORK ERROR, CONNECTION TIMED OUT. My all other instances works well.
I tried rebooting & resetting multiple times. Reading out stuff from internet since last 6 hours but for no luck. I also tried to clone disk of the instance and creating new instance with the cloned disk but for no luck. Same network connection issue. May be something in OS got corrupted? Please suggest. I have a number of websites hosted on the server which are down due to this. Thanks a lot in advance.
I took screenshot of my VM using Google Cloud Shell. It is as follows:
I connected with serial console which is as follows:
While creating a ticket with google, I found that they've posted some information under "Known Issues". I am pasting the whole stuff as there is no direct link to reach there. The symptoms they told were exactly what was happening to me:
Below is Known Issue Posted by Google:
Description:
We are experiencing an issue with Google Compute Engine instances running RHEL and CentOS 7 and 8. More details on this issue are available in the following article and bugs: https://access.redhat.com/solutions/5272311 https://bugzilla.redhat.com/show_bug.cgi?id=1861977 (RHEL 8) https://bugzilla.redhat.com/show_bug.cgi?id=1862045 (RHEL 7) Symptoms: Instances running RHEL and CentOS 7 and 8 that run yum update may fail to boot after restart with errors messages referring to a combination of: "X64 Exception Type - 0D(#GP - General Protection) CPU Apic ID", "FXSAVE_STATE" or "Find image based on IP". This issue affects instances with specific versions of the shim package installed. To find the currently installed shim version, use the following command: rpm -q shim-x64 Affected shim versions: CentOS 7: shim-x64-15-7.el7_9.x86_64 CentOS 8: shim-x64-15-13.el8.x86_64 RHEL 7: shim-x64-15-7.el7_8.x86_64 RHEL 8: shim-x64-15-14.el8_2.x86_64 Workaround: Do not update or reboot instances running RHEL or CentOS 7 and 8. If you are on an affected shim version, run yum downgrade shim\* grub2\* mokutil to downgrade to the correct version. This command may not work on CentOS 8. If you have already rebooted, you will need to attach the disk to another instance, chroot into the disk, then run the yum downgrade command. We will provide an update by Thursday, 2020-07-30 14:00 US/Pacific with current details.
Start time:
July 30, 2020 at 9:08:34 PM GMT+5
How to diagnose:
Instances running RHEL and CentOS 7 and 8 that run yum update may fail to boot after restart with errors messages referring to a combination of: "X64 Exception Type - 0D(#GP - General Protection) CPU Apic ID", "FXSAVE_STATE" or "Find image based on IP". This issue affects instances with specific versions of the shim package installed. To find the currently installed shim version, use the following command: rpm -q shim-x64 Affected shim versions: CentOS 7: shim-x64-15-7.el7_9.x86_64 CentOS 8: shim-x64-15-13.el8.x86_64 RHEL 7: shim-x64-15-7.el7_8.x86_64 RHEL 8: shim-x64-15-14.el8_2.x86_64
Workaround:
Do not update or reboot instances running RHEL or CentOS 7 and 8. If you are on an affected shim version, run yum downgrade shim\* grub2\* mokutil to downgrade to the correct version. This command may not work on CentOS 8. If you have already rebooted, you will need to attach the disk to another instance, chroot into the disk, then run the yum downgrade command.
Problem with starting minikube 1.1.1 on Ubuntu 19.04:
X Unable to start VM
* Error: [VBOX_HOST_ADAPTER] start: Error setting up host only network on machine start: The host-only adapter we just created is not visible. This is a well known VirtualBox bug. You might want to uninstall it and reinstall at least version 5.0.12 that is is supposed to fix this issue
I've seen this question: How to fix VM issue with minikube start ? however I'm running ubuntu 19.04, not OSX. I've also done Minikube not starting on Ubuntu, throwing errors a couple of times too.
Now, I've installed the dkms package, I've modprobed vbox-dkms into the kernel, UEFI secure boot is set up as part of the dkms installation, but I still can't get minikube to start
Anyone have any ideas where to look next or what the problem might be?
Ps. I'm running virtualbox 6.0.6_Ubuntu r129722
I have a problem with Docker running the nanoserver.
My environment: I Installed docker on a Win10 (developer build from microsoft) Virtual Machine (cause host is still Win7 with no default Docker support because of Hyper-V). I installed docker on the virgin image so no 3rd-party programs can cause the error. I also have already checked if Hyper-V is enabled.
But I think this Screenshot says everything:
Also Googled the problem, but everything i found wasn't in a Docker context.
FYI: I want to use the Docker container for running a network rendering slave which should only see one cpu core (for licensing reasons), maybe someone has another option for this.
so you are trying to run Docker on Windows 10 which is a VM on Windows 7? I suppose this is not possible. You are trying to run a virtualization platform inside a virtualized Host (your Windows 10 machine). This nested virtualization is not supported by Windows 7 afaik.
Nested virtualization is supported on Windows 10 Build 10565 and later (this must be your virtualization host).
Have you tried to create and run a Hyper-V VM inside that Windows 10 VM? this will also fail.
I have a CentOS 7.2 laptop. I have VirtualBox 5.0.x installed on it. Out of the box, it appears that I have Vagrant 1.7.2. Apparently this doesn't support VirtualBox 5.0.x, but it appears that version 1.8.1 does. What is the proper way to get access to 1.8.1? There is an installation rpm for 1.8.1 that I've downloaded, I can run this with "rpm", but do I have to do anything else before I do that?
Use Software Collections to get Vagrant 1.8.1 on CentOS: Vagrant 1.8 by Software Collections (the linked page contains instructions on how to do that). That will give you a build tested by the CentOS project.
As a sidenote: If you downloaded a standalone RPM package (presumably directly from the Vagrant project) and wish to install it, do it using the yum package manager, not rpm -- yum takes care of dependencies.
Just installed CentOS 5.8 on a VM using VMware Fusion 3.1.4 on my MB Mac OS X 10.7.4 and have the following problem. CentOS finished to install and re-start. The result, a nice black screen. I rebooted CentOS (on the VM) no success. I rebooted VMware Fusion, no success. I rebooted the whole machine (physical machine) no success.
Is there anywhere VMware Fusion log booting errors I can investigate why I have this annoying black screen?
VMware Fusion logs errors in the virtual machine's bundle.
So, for example, if your virtual machine appears on your hard disk on the desktop as:
myvm
you will find it in Terminal at:
cd ~/Desktop
ls -1 myvm*
... myvm.vmwarevm
then:
cd myvm.vmwarevm
you will see:
vmware.log
(and maybe vmware-0.log, etc, which will be older logs; vmware-0 being second most recent, vmware-1 being third most recent.)
A very late answer, but hope that is some help.