I'm new to podman, and I just trying to run containers on it.
(podman version 3.4.0, installed by brew, intel Core MAC)
However, when I trying to run "podman run {image-name}", below errors were thrown.
$ podman run -ti -d --name web httpd 125
Error: error configuring network namespace for container b0e70d672cb66005833c0a300c8661b88eab49e942c240d69d17587e0b75c47b: error adding pod web_2_web_2 to CNI network "podman": unexpected end of JSON input
$ podman run centos:7
Error: error preparing container a6d0bc1ad217cd8207935561dc8ff7bd33672da3fa513917f9965cb39520c449 for attach: error configuring network namespace for container a6d0bc1ad217cd8207935561dc8ff7bd33672da3fa513917f9965cb39520c449: error adding pod quirky_snyder_quirky_snyder to CNI network "podman": unexpected end of JSON input
By reading https://issueexplorer.com/issue/containers/podman/11452, I removed ~/.docker/, but the solution doesn't work in my case.
Of course, the error message says there was "unexpected end of JSON input", but I don't know how to fix it. Could anyone guess why podman didn't work even running these base images, or how to debug it?
Thanks in advance.
on macos, current machine version 3.3.1 has this problem. I had this problem on server version 3.3.1 and I do not encounter it on server version 3.4.0. You can check server version with podman version.
Try removing current machine and installing a newer one
podman machine stop
podman machine rm
podman machine init --image-path next
podman machine start
Check server version again with podman version.
Try running your image again.
This is the error i encountered when i updated my CentOS 8.1/RHEL 8.1 machines and all the KVMs are showing the error below:
error: internal error: process exited while connecting to monitor: 2020-06-09T12:41:10.410896Z qemu-kvm: -machine pc-q35-rhel8.1.0,accel=kvm,usb=off,vmport=off,smm=on,dump-guest-core=off: unsupported machine type
Use -machine help to list supported machines
Note: The problem states the machine type Q35 is not well stated/configured in your virtual Kernel based machines RUNNING on RHEL 8/ CentOS 8
[Step 1:] cat /etc/libvirt/qemu/*.xml | grep \<name'\| machine'
This will list the machine type in all of the KVMs installed.
[Output Snippet]
machine pc-q35-rhel8.1.0
[Step 2:] cd /etc/libvirt/qemu; ll
This will list all the xml files in connection with your KVMs
[Step 3:] At /etc/libvirt/qemu Use virsh edit <KVM file> ###Don't include .xml###
Navigate to machine
[Output Snippet]
<os>
<type arch='x86_64' machine='pc-q35-rhel8.1.0'>hvm</type>
<loader readonly='yes' secure='yes' type='pflash'>/usr/share/OVMF/OVMF_CODE.secboot.fd</loader>
<nvram>/var/lib/libvirt/qemu/nvram/Loadbalancer_VARS.fd</nvram>
<boot dev='hd'/>
</os>
Change machine='pc-q35-rhel8.1.0' to machine='q35'
shift + zz to save and quit
[Step 4:]
systemctl restart libvirtd && systemctl status -l libvirtd
virsh list --all
virsh start --domain <KVM>
Check the status of your running KVMs
virsh list --state-running
Now the issue should be resolved and your KVMs should be humming away.
Note though if head back in and check on the configuration xml file with virsh edit, you'll note that q35 converts to pc-q35-rhel7.6.0 automatically.
But this shouldn't be an issue.
Cheers :)
installed ejabberd от FreeBSD
then install mod_log_chat (ejabberdctl modules_update_specs -> ejabberdctl module_install mod_log_chat) and see error when starting ejabberd:
[error] Supervisor ejabberd_sup had child ejabberd_gen_mod_sup started with gen_mod:start_link() at undefined exit with reason {'EXIT',{undef,[{econf,any,[],[]},{mod_log_chat,mod_opt_type,1,[{file,"/root/.ejabberd-modules/sources/ejabberd-contrib/mod_log_chat/src/mod_log_chat.erl"},{line,282}]},{gen_mod,'-get_validators/2-fun-1-',3,[{file,"src/gen_mod.erl"},{line,512}]},{lists,foldl,3,[{file,"lists.erl"},{line,1263}]},{gen_mod,get_validators,2,[{file,"src/gen_mod.erl"},{line,499}]},{gen_mod,validate_opts,3,[{file,"src/gen_mod.erl"},{line,557}]},{gen_mod,start_module,5,[{file,"src/gen_mod.erl"},{line,209}]},{lists,foreach,...}]}} in context start_error
02:36:32.939 [critical] Failed to start ejabberd application: {error,{shutdown,{failed_to_start_child,ejabberd_gen_mod_sup,{'EXIT',{undef,[{econf,any,[],[]},{mod_log_chat,mod_opt_type,1,[{file,"/root/.ejabberd-modules/sources/ejabberd-contrib/mod_log_chat/src/mod_log_chat.erl"},{line,282}]},{gen_mod,'-get_validators/2-fun-1-',3,[{file,"src/gen_mod.erl"},{line,512}]},{lists,foldl,3,[{file,"lists.erl"},{line,1263}]},{gen_mod,get_validators,2,[{file,"src/gen_mod.erl"},{line,499}]},{gen_mod,validate_opts,3,[{file,"src/gen_mod.erl"},{line,557}]},{gen_mod,start_module,5,[{file,"src/gen_mod.erl"},{line,209}]},{lists,foreach,2,[{file,"lists.erl"},{line,1338}]}]}}}}}
Right, mod_log_chat requires ejabberd 19.08 or higher since this commit: https://github.com/processone/ejabberd-contrib/commit/857d350a7119d63fc85555cc66df7ff923b91e39
That was not documented in the module README.txt file...
You have two solutions:
Upgrade your ejabberd 19.05 to 19.08 or higher
or Downgrade your mod_log_chat source code to a version older than the mentioned one.
I met a issue when I install zabbix3.0 by packages on centos7.
When I done on mysql , php , apache and configuration in zabbix.conf.
I run systemctl start zabbix-server.service.It didn't work!and show :
Job for zabbix-server.service failed. See 'systemctl status zabbix-server.service' and 'journalctl -xn' for details.Then , my colleague told me to install trousers and gnutls,and then ,zabbix-server worked.What is the use of these two software?If they are necessary,Why not put them in the package of zabbix?
You won't start Zabbix Server 3.0 on CentOS 7 because you don't have "Disabled" SELinux.
You can disabled SELinux right here: /etc/selinux/config.
After that, you must reboot your server with reboot or shutdown -r now.
After reboot, confirm that the getenforce command returns Disabled.
Most likely, you didn't install those packages, but upgraded them. They are linked in through the Jabber/XMPP support.
This was a bug in RedHat packages that took some time to resolve, see this bugreport : https://bugzilla.redhat.com/show_bug.cgi?id=1071171
And this is the Zabbix issue tracking the same problem : https://support.zabbix.com/browse/ZBX-7790
I have KVM virtual machine running CentOS 7 as guest OS. I'm trying to attach an additional disk to it on the run (without shutting it down) using this command:
$ sudo virsh attach-disk centos --source /var/lib/libvirt/images/newdisk.img --target sdb --persistent
But receive an error:
error: Failed to attach disk
error: internal error: cannot update AppArmor profile 'libvirt-d2e7bbb8-c7b3-44ec-b0ea-27539e0df732'
If I do the same with Debian guest - everything is ok.
What is difference, how to solve that?
UPDATE:
I have a comment!
I compared two VM's xml and saw that CentOS have QEMU-agent in his configuration:
<channel type="unix">
<source mode="bind" path="/var/lib/libvirt/qemu/channel/target/centos_auto.org.qemu.guest_agent.0"></source>
<target name="org.qemu.guest_agent.0" type="virtio"></target>
<address bus="0" controller="0" port="1" type="virtio-serial"></address>
</channel>
Then I removed "channel qemu-ga", restarted VM and checked "hot add" feature. It worked.
I tested it on other VMs (CentOS, Fedora, Debian) and saw the same.
As a result:
If enable qemu-agent i cannot use hot plug.
If use "hot plug" i must forget about agent.
Is it my mistake in configuration or these features can't work together?
Host-OS: Ubuntu 15.10
QEMU emulator: now 2.4.92 (tested 2.3 and 2.4.1)
VMM: 1.3.0
This is a clear bug in the apparmor security driver for libvirt. The existence of the QEMU guest agent config in the XML should have no impact on ability to hotplug disks to a guest. This bug should be reported to the libvirt upstream, or Ubuntu bug trackers.