In order to debug a crash of VPP, it is required to provide a coredump file which allows backtracing of the VPP issue. The following are the requirements for capturing it:
As described in BUG_REPORTS.md.
In /etc/vpp/contiv-vswitch.conf
, add the following lines into the unix
section:
unix {
...
coredump-size unlimited
full-coredump
}
After (re)deploying Contiv-VPP networking, enter bash shell in the vswitch
container (use actual name of the vswitch POD - contiv-vswitch-7whk7
in this case):
kubectl exec -it contiv-vswitch-7whk7 -n kube-system -c contiv-vswitch bash
Enable coredumps:
mkdir -p /tmp/dumps
sysctl -w debug.exception-trace=1
sysctl -w kernel.core_pattern="/tmp/dumps/%e-%t"
ulimit -c unlimited
echo 2 > /proc/sys/fs/suid_dumpable
Now repeat the steps that lead to VPP crash. You can also force VPP to crash at the point where it is running (e.g. if it is stuck) by sending it the SIGQUIT signal:
kill -3 `pidof vpp`
The core file should appear in /tmp/dumps
in the container:
cd /tmp/dumps
ls
vpp_main-1524124440
You can try to backtrace it after installing gdb:
apt-get update && apt-get install gdb
gdb vpp vpp_main-1524124440
(gdb) bt
Finally, copy the core file out of the container. First, still inside of the container, pack the core file into an archive:
cd /tmp/dumps
tar cvzf vppdump.tar.gz vpp_main-1524124440
Now, on the host, determine docker ID of the container and copy the file out of it:
docker ps | grep vswitch_contiv
d7aceb2e4876 c43a70ac3d01 "/usr/bin/supervisor…" 25 minutes ago Up 25 minutes k8s_contiv-vswitch_contiv-vswitch-zqzn6_kube-system_9923952f-43a6-11e8-be84-080027de08ea_0
docker cp d7aceb2e4876:/tmp/dumps/vppdump.tar.gz .
Now you are ready to file a bug in jira.fd.io and attach the core file.