Problem with migrating vms after host error hook launched

  1. It seems that there is a bug that causes the VM to change to running state when disk monitoring arrives. I’ve opened an issue:

https://dev.opennebula.org/issues/5331

  1. EXECUTE SUCCESS 1 error: means that the hook was successfully executed and wrote no error message

  2. The VM changed from UNKNOWN to RUNNING before the scheduler picked it for rescheduling. That made it think it’s really running and tried to do a live migration. On UNKNOWN state the VM is not migrated, it’s redeployed. This was caused by the bug you’ve found.

  3. We aim to fix the bug in version 5.4.1. Meanwhile, as a workaround, you can disable disk monitoring. To do this you can comment the following lines from /var/lib/one/remotes/tm/qcow2/monitor and issue the command onehost sync --force in the frontend:

1 Like