Hi I’m getting this message to some containers that are running. I’ve rebooted the containers and they reboot fine with mysql (percona) running on them. The waiting to bootstrap message is misleading; how can I fix this?
juju version: 2.8.8-bionic-amd64
cadmin@conductor:~$ juju status | ag mysql
mysql 5.7.20-29.24 waiting 3 percona-cluster jujucharms 269 ubuntu
mysql/0* active idle 0/lxd/7 10.2.2.51 3306/tcp Unit is ready
mysql/1 waiting idle 15/lxd/1 10.2.2.217 3306/tcp Unit waiting to bootstrap
mysql/2 waiting idle 16/lxd/1 10.2.2.218 3306/tcp Unit waiting to bootstrap
If the containers think that they’re still “waiting to bootstrap”, it means that they haven’t checked in to tell the controller that they’re up and running for some reason.
Please take a look at the juju agent logs on those containers (in /var/log/juju) and let us know if you see any interesting errors. I suspect that either the controller ip has changed, and the agents can’t find it as a result, or there is some other issue on the machine that is knocking the juju agent over before it can tell the controller that it is up and running and has done the work that the controller expected it to do.
Lxd recently added support for virtual machines, in addition to containers. It looks like it’s initially trying to spin up a kvm (kernel virtual machine), and then failing. I’m not sure why it would work on restart.
There are two possibilities:
We have a bug in Juju, where it is not correctly asking lxd for containers.
You configured your container cloud to use kvm by default, and that is causing trouble, because your host machine isn’t configured to support virtualization.
If option #2 doesn’t sound likely, let me know. I’ll look into option #1 in the meantime. (There were some app armor changes in focal that broke the way that we were testing lxd – it’s possible that when we refactored out tests, we dropped in some non default configuration for lxd, and missed that this could be an issue.)
Thank you for looking into this. We found that there was a missing file on some of the machines called “seeded” in the percona directory. That solved the reporting it seems. I hope this helps someone.