Running KVM in addition to LXC on local LXD CLOUD --SOLVED (by using LTS kernel [4.8] for Xenial 16.04.2)

N. S. n5pascal at hotmail.com
Wed Jun 28 22:59:34 UTC 2017


Hi again,

Sorry for late message.

Below problem is solved by:
Killing the controller and installing the Lts enablement stack as pointed out by Nicholas on below link.
Thank you Nicholas...

Such kernel did include the zfs module which is natively supported in Xenial.

Using the enablement stack, this has installed generic 4.8 kernel with zfs dkms and all its dependencies.

Low latency kernel is then installed on top of it by
$sudo dkpg -i Linux-headers-<ver>-lowlatency Linux-image-<ver>-lowlatency


Thanks again,

BR,
NS
On Jun 28, 2017 8:15 PM, "N. S." <n5pascal at hotmail.com> wrote:

Hi Nicholas and Andrew,


Thank you both for your help.

@Nicholas,
I downloaded the kernel 4.8 and 4.11 Low-latency (Requirement by the application) and installed it
sudo dpkg -i linux-header-4.8* linux-image-4.8*
sudo update-grub

But I am faced with a challenge that ZFS is not starting UP during boot. Cf attached.

Further digging, the status of the services is as follows:



$ systemctl status zfs-import-cache.service
● zfs-import-cache.service - Import ZFS pools by cache file
   Loaded: loaded (/lib/systemd/system/zfs-import-cache.service; static; vendor preset: enabled)
   Active: failed (Result: exit-code) since Wed 2017-06-28 19:54:33 EEST; 12min ago
  Process: 1032 ExecStartPre=/sbin/modprobe zfs (code=exited, status=1/FAILURE)

Jun 28 19:54:33 ns-HP systemd[1]: Starting Import ZFS pools by cache file...
Jun 28 19:54:33 ns-HP modprobe[1032]: modprobe: FATAL: Module zfs not found in directory /lib/modules/4.8.0-040800-lowlatency
Jun 28 19:54:33 ns-HP systemd[1]: zfs-import-cache.service: Control process exited, code=exited status=1
Jun 28 19:54:33 ns-HP systemd[1]: Failed to start Import ZFS pools by cache file.
Jun 28 19:54:33 ns-HP systemd[1]: zfs-import-cache.service: Unit entered failed state.
Jun 28 19:54:33 ns-HP systemd[1]: zfs-import-cache.service: Failed with result 'exit-code'.




ns at ns-HP:/usr/src/linux-headers-4.4.0-81$ systemctl status zfs-mount.service
● zfs-mount.service - Mount ZFS filesystems
   Loaded: loaded (/lib/systemd/system/zfs-mount.service; static; vendor preset: enabled)
   Active: failed (Result: exit-code) since Wed 2017-06-28 19:54:34 EEST; 13min ago
  Process: 1042 ExecStart=/sbin/zfs mount -a (code=exited, status=1/FAILURE)
 Main PID: 1042 (code=exited, status=1/FAILURE)

Jun 28 19:54:33 ns-HP systemd[1]: Starting Mount ZFS filesystems...
Jun 28 19:54:34 ns-HP zfs[1042]: The ZFS modules are not loaded.
Jun 28 19:54:34 ns-HP zfs[1042]: Try running '/sbin/modprobe zfs' as root to load them.
Jun 28 19:54:34 ns-HP systemd[1]: zfs-mount.service: Main process exited, code=exited, status=1/FAILURE
Jun 28 19:54:34 ns-HP systemd[1]: Failed to start Mount ZFS filesystems.
Jun 28 19:54:34 ns-HP systemd[1]: zfs-mount.service: Unit entered failed state.
Jun 28 19:54:34 ns-HP systemd[1]: zfs-mount.service: Failed with result 'exit-code'.
ns at ns-HP:/usr/src/linux-headers-4.4.0-81$

I tried to run modprobe zfs as advised, but:

$ sudo -i
[sudo] password for ns:
root at ns-HP:~# /sbin/modprobe zfs
modprobe: FATAL: Module zfs not found in directory /lib/modules/4.8.0-040800-lowlatency
root at ns-HP:~#



I know this might be not directly related to JUJU, but to ubuntu kernel, but I appreciate if  you could help.


Thanks,

BR,

NS



Nicholas Skaggsnicholas.skaggs at canonical.com<mailto:juju%40lists.ubuntu.com?Subject=Re%3A%20Running%20KVM%20in%20addition%20to%20LXC%20on%20local%20LXD%20CLOUD&In-Reply-To=%3CCAHNOz3oXQF4t2x5Rz6R5P-zfuQ651PPJSJEQRf9Er-ZW_3pE4Q%40mail.gmail.com%3E>
Tue Jun 27 19:51:08 UTC 2017

If it's possible, I would simply run the hwe kernel on xenial which
provides 4.8+. Read more about running an updated stack here:

https://wiki.ubuntu.com/Kernel/LTSEnablementStack

This would solve your specific problem without worrying about running
kvm's.

Kernel/LTSEnablementStack - Ubuntu Wiki<https://wiki.ubuntu.com/Kernel/LTSEnablementStack>
wiki.ubuntu.com
Ubuntu Kernel Release Schedule. The following is a generic view of the Ubuntu release schedule, the kernels delivered, and the support time frames.





________________________________
From: Andrew Wilkins <andrew.wilkins at canonical.com>
Sent: Sunday, June 25, 2017 10:42 PM
To: N. S.; juju at lists.ubuntu.com
Subject: Re: Running KVM in addition to LXC on local LXD CLOUD

On Sat, Jun 24, 2017 at 9:14 PM N. S. <n5pascal at hotmail.com<mailto:n5pascal at hotmail.com>> wrote:
Hi,


I am running 10 machines on local LXD cloud, and it's fine.

My host is Ubuntu 16.04, kernel 4.4.0-81.

However, I have the following challenge:
One of the machines (M0) stipulates a kernel 4.7+


As it's known, unlike KVM, LXC makes use of same kernel of the host system and in this case (4.4.0-81) breaching thus the stipulation of M0 (4.7+).


I have read that starting Juju 2.0, KVM is no more supported.

Juju still supports kvm, but the old "local" provider which supported lxc/kvm is gone.

You could run a kvm container from within a lxd machine with the right apparmor settings. Probably the most straight forward thing to do, though, would be to create a KVM VM yourself, install Ubuntu on it, and then manually provision it using "juju add-machine ssh:<hostname>".


How could I prepare the stipulation of M0?

Thanks for your help
BR,
Nazih

--
Juju mailing list
Juju at lists.ubuntu.com<mailto:Juju at lists.ubuntu.com>
Modify settings or unsubscribe at: https://lists.ubuntu.com/mailman/listinfo/juju

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <https://lists.ubuntu.com/archives/juju/attachments/20170628/d5347a34/attachment-0001.html>


More information about the Juju mailing list