Containers like LXC are becoming a popular solution to program isolation. Compared to virtual machines (VM), containers tend to have less resource overhead and higher performance, which makes it interesting to explore how much benefit we can get from deploying virtual network functions (VNF) with containers instead of VMs. Therefore, we conducted an experiment in which we compared performance and resource usage of Suricata, a popular multi-threaded IDS program, in bare metal, Docker container, and virtual machine setups, and in different load levels and resource allocation configurations.
Note: the experiment was conducted during April - May 2016.
Code repository: GitHub
The full text is accessible here.
The huge overhead showed in VM is largely due to profiling (which relies
heavily on X86
rdtsc instruction) being enabled.
Use tcpreplay to replay some Pcap traffic files collected from the Internet, and analyze performance of Suricata (running in bare metal, Docker, and VM, respectively) from statistics it reports and resource usage of related processes and resource availability of the entire host.
When comparing Docker container and VM, we will also tune the resource limit and see how Suricata will perform.
We use four servers (the cluster of which is called “cap”) of the same hardware configuration. They form two pairs of test setups – (cap03, cap09) and (cap06, cap07). cap03 and cap06 are hosts to run Suricata, while cap09 and cap07 are hosts to run tcpreplay (which is CPU intensive) and control script. The two hosts of each pair are connected directly by an Ethernet cable.
All four machines have the following hardware configuration:
- CPU: Intel Xeon X3430 @ 2.40GHz (Nehalem; EIST/C-states disabled; VT-x on, VT-d on; HT unsupported)
- RAM: 2 x 2 GB DDR3-1333 RAM
- HDD: 500GB Seagate 3.5” 7200RPM + 2 x 1TB Seagate 3.5” 7200RPM
- Network: 2 x Broadcom 1Gbps NIC. em1 (enp32s0 on cap03/cap06) is used for remote access and management, and em2 (enp34s0 on cap03/cap06) is used to transmit the test traffic.
Sender Host (cap07/cap09)
Senders run Ubuntu Server 14.04.4 64-bit.
Receiver Host (cap03/cap06)
Receivers run Ubuntu Server 15.10 64-bit. The reason is that many packages (particularly QEMU, which dates 8 years ago) on Ubuntu 14.04 are too old; some (libvirt 1.2.2) are even buggy.
The docker image and VM try to resemble software of host system, running Ubuntu Server 15.10 64-bit and making sure all critical packages are of the same version as the host.
- To use streamlined test script, host user and sender must be able to run sudo without password prompt
sudo visudoand add
username ALL=(ALL) NOPASSWD: ALL).
- Use SSH authorized_keys to facilitate SSH login.
Suricata loads the free version of Emerging Rules as of 2016-04-14.
Unless otherwise noted, Suricata uses default configuration generated when installed to the system.
Suricata 3.0.1, released on April 4, 2016, fixed many memory leak bugs and improved stability. This can be confirmed by our previous testing of 3.0 version inside VM setup, which resulted in memory thrashing and can barely be tested in any load above moderate.
We use the following flows available from the Internet:
- Sample flows provided by TCPreplay –
- Sample traces collected by WireShark
- Publicly available PCAP files
- ISTS’12 trace files – randomly picked
According to TCPreplay website, bigFlows.pcap has the following characteristics:
This is a capture of real network traffic on a busy private network’s access point to the Internet. The capture is much larger and has a smaller average packet size than the previous capture. It also has many more flows and different applications. If the large size of this file isn’t a problem, you may want to select it for your tests.
- Size: 368 MB
- Packets: 791615
- Flows: 40686
- Average packet size: 449 bytes
- Duration: 5 minutes
- Number Applications: 132
From the log of Suricata we can confirm that the traffic consists of numerous procotols on various ISO/OSI layers.
The peak throughput of the trace is approximately 2.13 MBps (diagram will given later), so our Ethernet link will support less than 50 concurrent TCPreplay processes. However, CPU of the sender host saturates at 4 concurrent TCPreplay processes; it takes longer for all TCPreplay processes to finish replaying above that concurrency level (e.g., peak throughput less than doubles at concurrency level 16 compared to concurrency level 4, because at level 16 it can take 3X more time to finish). For this reason, test results above concurrency level 4 are not comparable to those equal to or below that level. Results not in the same concurrency level above 4 are not cross-comparable either.
This trace file requires much higher processing speed than
- Rated: 6928500.0 Bps, 55.42 Mbps, 6264.04 pps
- Flows: 1833 flows, 80.74 fps, 140196 flow packets, 2006 non-flow
- Duration: ~22 seconds
We analyze the performance of Suricata by comparing speed of packets / bytes captured, speed of packets / bytes decoded, and speed of alerts triggered for different setups when playing the same trace with the same number of parallel workers. Theoretically the speed of traffic sent is the same, so the difference of those speeds result from the receiver setup.
Suricata generates stats every 8 seconds (default value).
We use a Python script based on Python3
psutil package. It periodically polls the system-wide CPU, memory, and Disk I/O usage, and
optionally the traffic sent / received on specified NIC, and sum of resource usage of a subset of processes (e.g.,
and their children processes) and outputs the information to a CSV file. We compared the numbers with popular system monitor programs
atop (buggy, discussed later) and can confirm that the numbers are correct and the periods between polls are
(almost perfectly) constant.
Resource monitor polls stats every 4 seconds.
The source code of resource monitor is hosted in
Why not use
atoppolls more numbers than needed, they incur much higher overhead on the receiver system.
- Output is not friendly for post-processing, even after CSV-fied.
- Lines of
topdo not contain timestamps; because we start monitors before Suricata, we cannot ask monitors to poll information of Suricata process (its PID not known beforehand).
atopbehaves wrong on our system. When running VM setup,
atopreports 50% of CPU usage each core while in fact they are all 100% (confirmed by
top(which needs to monitor several processes like two
dockerprocesses and one
Suricata-Mainprocess) are separate processes, their output seldom hits the same timestamp, and thus the sum of the reported numbers is not accurate.
By contrast, our monitor script prints neatly for each polling timestamp all the system-wide resource availability and sum of resource usage of the processes we are interested in one CSV line. This makes post-processing easy as well.
Why examine system-wide resource usage?
Because the sum of resource usage of processes that are directly related may not be comprehensive. For example, the RAM usage of mirroring traffic to macvtap is not in any suricata, docker, or qemu processes.
System-wide numbers could over-count, but as we are comparing the difference, which results from different setups of Suricata, the over-counted part doesn’t matter.
In bare metal setting, Suricata will run directly on top of hardware and inspect the NIC that the test traffic enters.
In this Docker setting, the container is configured so that the network interfaces of the host is exposed to the container, enabling Suricata to inspect the same NIC interface as in bare metal setting.
The CPU and RAM limitations can be passed as parameters of the test script. By default, it allows the container to access all 4 cores and has a RAM limit of 2GB (1536m and 1g are also tested).
Docker + macvtap
In Docker-vtap setting, the difference from Docker setup is that we create a macvtap of model “virtio” and mode “passthrough” to mirror the traffic arriving at host’s enp34s0 NIC, and let Suricata in Docker inspect the traffic on the macvtap device. This is an “intermediate” setup between Docker setup and VM setup, because the cost of running macvtap is unavoidable in VM setup.
In the text that follows, I will use “DockerV” to refer to this setup.
The virtual machine hardware is configurable. Default configurations are available in XML format in
config/. Different CPU and RAM configuration may be passed as parameters of the test script.
By default, it has 4 vCPUs each of which has access to all 4 cores of the host CPU. Capacities of vCPU are copied from host CPU
(“host-model”). RAM is set to 2 GB. In terms of NIC, We create a macvtap device (macvtap0) of model “virtio” and mode “passthrough” to
copy the test traffic arriving at host’s enp34s0 to VM’s eth1. Another NIC, eth0, is added for communications between host and the VM.
The exact hardware configuration will be mentioned when comparing results.
The virtual disk has size of 64 GiB, large enough to hold logs of GB magnitude.
We have the following tests:
|Setup||Trace file||Para. TCPreplays||Use VTAP?||Memory*||CPU||Swappiness||Other Args||Sample Size|
|Bare metal||bigFlows.pcap||1||No||4 GB||4||5||-||40|
|Docker + vtap||bigFlows.pcap||1||Yes||2 GB||4||5||-||40|
|Bare metal||bigFlows.pcap||2||No||4 GB||4||5||-||40|
|Docker + vtap||bigFlows.pcap||2||Yes||2 GB||4||5||-||40|
|Bare metal||bigFlows.pcap||3||No||4 GB||4||5||-||40|
|Docker + vtap||bigFlows.pcap||3||Yes||2 GB||4||5||-||40|
|Bare metal||bigFlows.pcap||4||No||4 GB||4||5||-||40|
|Docker + vtap||bigFlows.pcap||4||Yes||2 GB||4||5||-||40|
|Bare metal||bigFlows.pcap||8||No||4 GB||4||5||-||30|
|Docker + vtap||bigFlows.pcap||8||Yes||2 GB||4||5||-||25|
|Bare metal||bigFlows.pcap||16||No||4 GB||4||5||-||10|
|Docker + vtap||bigFlows.pcap||16||Yes||2 GB||4||5||-||10|
|Docker + vtap||bigFlows.pcap||4||Yes||1024 MB||4||5||-||30|
|Docker + vtap||bigFlows.pcap||4||Yes||512 MB||4||5||-||10|
|Bare metal||snort.log.1425823194||1||No||4 GB||4||5||stat=1s||30|
|Docker + vtap||snort.log.1425823194||1||Yes||2 GB||4||5||stat=1s||30|
|VM||snort.log.1425823194||1||Yes||2 GB||4||5||stat=1s, vCPUs=4||30|
|Bare metal||snort.log.1425823194||2||No||4 GB||4||5||stat=1s||30|
|Docker + vtap||snort.log.1425823194||2||Yes||2 GB||4||5||stat=1s||30|
|VM||snort.log.1425823194||2||Yes||2 GB||4||5||stat=1s, vCPUs=4||30|
|Bare metal||snort.log.1425823194||4||No||4 GB||4||5||stat=1s||30|
|Docker + vtap||snort.log.1425823194||4||Yes||2 GB||4||5||stat=1s||30|
|VM||snort.log.1425823194||4||Yes||2 GB||4||5||stat=1s, vCPUs=4||30|
From the perspective of physical host, neither Docker nor QEMU strictly enforces this memory limit. The actual memory usage can go slightly higher.
For some test setups, the variation is so low that it’s not worth repeating many times. For some tests the variation is high and more rounds are run to get satisfiably large sample. Some tests were run significantly more times because the test scripts were copied and pasted.
The number of CPU cores is not manipulated because we can gain insight from different load levels. Actually CPU is the bottleneck for most test setups.
We ran each test multiple times to generate a number of samples (as shown the sample size column). Before running every test, the receiver host is rebooted to make sure system state is restored back to original. After all samples are generated, we use the median of all samples at each checkpoint to obtain an average result of the test. We then compare the average results of tests.
The script to parse raw CSV / Suricata eve.json files is hosted in
Note: after a test is run, all generated log files will be transferred to server cap08 using
rsync. There are occasional glitches on
public network which could cause failure (connection timeout) of transmission. This is the reason why the sample size of some tests are
slightly smaller. I don’t know why the clusters become inaccessible during those minutes.
We analyze the results of the two trace files, respectively.
The following is the recorded throughput of trace
bigFlows.pcap when played by one TCPreplay process. When there are two, three, or four TCPreplay processes, the throughput is simply multiplied because the sender host is not saturated in terms of CPU, memory, or NIC throughput. This can be confirmed by the system stat log of sender host.
In the following sections, I’ll use “load” to mean a unit of TCPreplay process. For example, “2X load” means using two concurrent TCPreplay processes.
Comparing CPU Usage of Four Setups
It takes about 8 seconds for Suricata to initialize for all four setups. After Suricata loads, the sender will start replaying the trace.
At 1X load, we see that Docker and DockerV setups use almost the same amount of CPU share as that of bare metal setup, fluctuating near 25%, whereas the VM raises the host CPU usage to a range between 200% and 350%. The CPU usage of VM setup fluctuates greatly but is highly correspondent to the trace throughput. Note that when the VM runs without Suricata, the host CPU usage is about 2% to 5%.
At 2X load level, we see that the CPU usage of bare metal, Docker, and DockerV setups doubles but still close to each other, and the VM setup has host CPU saturated.
3X and 4X load levels reveal similar result – CPU usage tripled and quadrupled, respectively, compared to 1X load level, and VM setup consumed all host CPU resource.
Comparing CPU Usage of Bare Metal and Docker Setups
Here is a diagram that puts the CPU usage of all setups except for VM and at ll load levels:
We see that Docker layer imposes trivial (with respect to overall CPU usage) overhead to host CPU when Suricata inspects the trace traffic at all four load levels.
But what’s the overhead of macvtap traffic mirroring?
CPU Overhead of Docker and Macvtap
By comparing the CPU usage of Bare metal, Docker, and DockerV setups, we can gain insight on how much overhead Docker and macvtap introduces, respectively:
While the CPU usage overhead of Docker is trivial at 4X load, there is a roughly 0% to 5% increase in CPU usage with macvtap, depending on traffic throughput. Note that we found macvtap of mode “passthrough” (requiring VT-d and SR-IOV) and model “virtio” the approach that incurs the least overhead compared to other ways of redirecting traffic including bridging, Virtual Ethernet Port Aggregator (VEPA), etc. With other approaches of redirecting traffic, the overhead can go higher.
By making this comparison we see that
- By exposing the host NIC directly to Docker container, the overhead of mirroring traffic can be saved.
- The high CPU usage of VM setup is not caused by macvtap traffic redirection; however, the CPU resource needed to redirect traffic it not negligible.
Comparing Memory Usage of Four Setups
The result of memory usage is so simple that they can be put in one diagram:
We see that increasing the load level barely increases the host memory usage on bare metal, Docker, and Docker with vtap setups. However, increasing load level dramatically increases the host memory usage on VM setup.
Note that even though after 150 seconds the memory usage of 2X, 3X, 4X load levels is about the same for VM setups, for those load levels, it’s the host CPU that is a bottleneck. This prevents Suricata from increasing speed and taking more memory, and also explains why the memory usage there slightly decreases as load increases – CPU is saturated and the system needs more CPU time to handle the network traffic, making Suricata run with even less resource. It’s still noteworthy that for VM setup when CPU is not a bottleneck, the host observes an up to 15% memory usage increase. I therefore conjecture that if CPU were not a bottleneck, the memory usage will go spectacularly higher.
In terms of overhead, while the bare metal consumes about 10% of CPU usage, Docker setups impose trivial memory head, whereas the VM setup can eat three times as much, resulting in a up to 250% overhead when busy and 300% when idle.
Memory Usage inside KVM
We also investigate how much memory is used inside VM and on host. From the graph below we see that the VM has not saturated its 2GB memory limit. We therefore understand that for our test setups memory is not a bottleneck.
Comparing Performance of Suricata in Four Setups
Suricata exports performance metric in intervals of, by default, 8 seconds, to log files. We examine the log file and compare the following metrics:
- _capture.kernelpackets: The accumulative number of packets Suricata has captured.
- _capture.kerneldrops: The accumulative number of packets Suricata has dropped.
- decoder.pkts: The accumulative number of packets Suricata has decoded.
- decoder.bytes: The accumulative amount of data, in Bytes, Suricata has decoded. We post processed the data to use unit of KiB rather than B.
We do not examine the number of alerts triggered because it’s highly affected by packet dropping.
We see that in all four load levels we use, packet capturing is about the same for all four setups. However, the VM setup tends to receive less packets in the end. Usually TCPreplay ends at second 312 and we send SIGTERM signal to Suricata 30 seconds after and wait for it to exit gracefully. It’s very likely that for VM setup it still has packets yet to capture when exiting, resulting in the discrepancy we see here.
Packet dropping is a sign that Suricata can’t process the workload given the resource it can utilize. Only VM setup observes packet drop starting from 2X load, and almost all increased load beyond 2X is dropped.
Because the data cannot fit well in one graph, we first compare Docker setups with bare metal, and then VM setup with baremetal.
We first compare Docker and DockerV setups with bare metal, and see that there is no nontrivial difference between the three. In this case, there is no need to compare the number of bytes decoded because they will be on par with each other.
We then compare VM setup with bare metal.
We see that while the decoding speed is comparable at 1X load, the decoding speed of VM Suricata is significantly lower as load increases from 2X to 4X, and there is no difference in decoding between 3X load and 4X load, which indicates that Suricata is saturated.
If we switch unit to KBytes, we see that as load increases, Suricata in VM runs slower and slower.
When Memory Is Also a Bottleneck
In the previous tests, we set memory limit of 2GB which turns out sufficient for all setups and all four load levels. In this part we change memory limit to 512MB and see how resource usage and performance would change. From our previous tests we see that even 512MB is sufficient for 1X load on VM yet CPU becomes bottleneck of VM since 2X load. While we could tweak the parameters to isolate the CPU bottleneck, it’s also reasonable to change the memory limit and observe the difference. Note that we did test with 1GB memory limit, but it turns out no difference from 2GB limit since VM’s CPU bottleneck prevented VM from consuming more memory.
Therefore we compare the following two setups: DockerV versus VM under 512MB memory limit and 4X load. We choose those two setups so that the overhead of macvtap appears on both sides. For convenience we include their 2GB counterparts in the graphs.
Host CPU Usage
We see that 512MB has no effect on DockerV setup, but the CPU usage starts to decrease for 512MB VM since around 150 second. As we will see in a later graph, VM memory saturates there and thrashing occurs.
Host Memory Usage
There is not much to say about host memory usage. There is no difference between 2GB limit and 512MB limit in DockerV setup, while the VM memory hits its limit and cannot increase further. The memory increase after the point of thrashing is largely due to host caching the disk swap I/O of VM.
Memory Usage in VM
Memory usage in VM confirms that thrashing starts after approx. 150 seconds. We will inspect the memory usage drop near 328 second with behavior of Suricata.
Performance of Suricata
We see that Suricata in DockerV setup shows no difference in performance given the two memory limits as the lines for their packet captured match well, and neither drop any packets. KVM with 2GB memory limit matches the DockerV packet captured line well until near the end; its lack of CPU power makes it impossible for Suricata to finish with all packets within the given time frame.
What’s worth noting is that the graph seems to indicate that 512MB KVM Suricata captures more packets than other scenarios. However, this is not the case. Memory thrashing inside VM drifts Suricata’s logging intervals (in unit of seconds), and when we put all series into one X-axis the drift is assumed non-existent even though it gets increasingly severe as time goes by. This can be verified by comparing the raw statistics of NIC traffic on host and inside VM manually.
We see that Suricata inside 512MB VM is struggling to make slow progress, while Suricata in 512MB DockerV works just fine. Note that the slight drop of the line for 512MB KVM near 360 second is caused by the fact that we take medians of samples.
Increasing the Load Level
From result of other test configurations, we see that Docker setup is on a par with bare metal in terms of both resource usage and performance, up to 32X load level, beyond which we did not test since the sender host would be overly saturated.
Why is Suricata in VM that slow?
To further investigate how Suricata consumes way more CPU in VM than in bare metal, we use
oprofile to profile the execution of Suricata with 4X load in bare metal and in VM, respectively.
The result is hosted in
The function calls that consume most of (aggregated) CPU time on host execution are:
samples % linenr info app name symbol name 3079940 48.7275 util-cpu.c:190 suricata UtilCpuGetTicks 889800 14.0775 util-mpm-ac.c:1307 suricata SCACSearch 573918 9.0799 detect.c:1237 suricata SigMatchSignatures
While the function calls that consume most of CPU time on VM execution are:
samples % linenr info app name symbol name 3876496 77.4428 util-cpu.c:190 suricata UtilCpuGetTicks 378915 7.5698 detect.c:1237 suricata SigMatchSignatures 144521 2.8872 util-mpm-ac.c:1307 suricata SCACSearch 78725 1.5727 util-profiling.c:994 suricata SCProfileRuleStart
We see that the function
UtilCpuGetTicks() consumes about 48.73% of CPU time of Suricata process on bare metal whereas it consumes
77.44% in KVM. With detailed report from oprofile we see that it is the seven processing threads that rely heavily on this function,
whereas the four management threads do not.
We then wonder which instruction in this function takes most time. Assembly wise, on host
0000000000736ca0 <UtilCpuGetTicks>: /* UtilCpuGetTicks total: 3079940 48.7275 */ 19614 0.3103 : 736ca0: push %rbx 1345 0.0213 : 736ca1: xor %eax,%eax : 736ca3: cpuid 1281367 20.2724 : 736ca5: rdtsc 391861 6.1996 : 736ca7: mov %edx,%edi 20 3.2e-04 : 736ca9: mov %eax,%esi : 736cab: xor %eax,%eax : 736cad: cpuid 1300635 20.5772 : 736caf: mov %rdi,%rax 33929 0.5368 : 736cb2: pop %rbx 51165 0.8095 : 736cb3: shl $0x20,%rax 4 6.3e-05 : 736cb7: or %rsi,%rax : 736cba: retq : 736cbb: nopl 0x0(%rax,%rax,1)
while in KVM
0000000000736ca0 <UtilCpuGetTicks>: /* UtilCpuGetTicks total: 3876496 77.4428 */ 3201 0.0639 : 736ca0: push %rbx 421 0.0084 : 736ca1: xor %eax,%eax 1 2.0e-05 : 736ca3: cpuid 1760507 35.1706 : 736ca5: rdtsc 159997 3.1963 : 736ca7: mov %edx,%edi 5 1.0e-04 : 736ca9: mov %eax,%esi : 736cab: xor %eax,%eax 2477 0.0495 : 736cad: cpuid 1783625 35.6324 : 736caf: mov %rdi,%rax 110918 2.2159 : 736cb2: pop %rbx 55341 1.1056 : 736cb3: shl $0x20,%rax 3 6.0e-05 : 736cb7: or %rsi,%rax : 736cba: retq : 736cbb: nopl 0x0(%rax,%rax,1)
We find that it takes significantly more (aggregated) time to read x86 Timestamp Counter (TSC) register as the instruction
shows, and to move register value from
Unfortunately we fail to obtain a callgraph from oprofile which could let us know the reason why it spends more time on those instructions – is it simply because those instructions run slowly in VM? Or is it because this function is triggered more frequently?
To investigate this, we copy the source code of
UtilCpuGetTicks into separate C file test_rtdsc.c, and write a
main function that calls this function 10^8 times. We run this program in both bare metal and VM, measuring how much time it takes to finish with
time command, and the distribution of time onto the instructions with oprofile. The raw result is saved in test_rdtsc/ directory.
We find that it takes only 6.6 seconds (43.35% attributable to
rdtsc) in bare metal, but 126.09 seconds (47.90% attributable to
rdtsc) inside VM. We also note that the instruction to update the loop counter, which is the first instruction after
takes significant amount of time (44.34% in bare metal; 47.44% in VM).
According to Table 21-6, Intel Manual Volume 3B, pp.21-13, the instruction
rdtsc could trigger a VM exit, making the instruction expensive in VM environment.
cpuid is often paired with
rdtsc to prevent
rdtsc from being executed out-of-order:
In order to keep the RDTSC instruction from being performed out-of-order, a serializing instruction is required. A serializing instruction will force every preceding instruction in the code to complete before allowing the program to continue. One such instruction is the CPUID instruction, which is normally used to identify the processor on which the program is being run. For the purposes of this paper, the CPUID instruction will only be used to force the in-order execution of the RDTSC instruction.
With a tool provided in QEMU repository written by Avi Kivity [email protected], we can check the “Primary Processor-Based VM-Execution Controls” of our test CPU:
sudo ./vmxcap.py Basic VMX Information Revision 14 VMCS size 1024 VMCS restricted to 32 bit addresses no Dual-monitor support yes VMCS memory type 6 INS/OUTS instruction information yes IA32_VMX_TRUE_*_CTLS support yes pin-based controls External interrupt exiting yes NMI exiting yes Virtual NMIs yes Activate VMX-preemption timer yes Process posted interrupts no primary processor-based controls Interrupt window exiting yes Use TSC offsetting yes HLT exiting yes INVLPG exiting yes MWAIT exiting yes RDPMC exiting yes RDTSC exiting yes CR3-load exiting default CR3-store exiting default CR8-load exiting yes CR8-store exiting yes Use TPR shadow yes NMI-window exiting yes MOV-DR exiting yes Unconditional I/O exiting yes Use I/O bitmaps yes Monitor trap flag yes Use MSR bitmaps yes MONITOR exiting yes PAUSE exiting yes Activate secondary control yes secondary processor-based controls Virtualize APIC accesses yes Enable EPT yes Descriptor-table exiting yes Enable RDTSCP yes Virtualize x2APIC mode yes Enable VPID yes WBINVD exiting yes Unrestricted guest no APIC register emulation no Virtual interrupt delivery no PAUSE-loop exiting no RDRAND exiting no Enable INVPCID no Enable VM functions no VMCS shadowing no RDSEED exiting no EPT-violation #VE no Enable XSAVES/XRSTORS no VM-Exit controls Save debug controls default Host address-space size yes Load IA32_PERF_GLOBAL_CTRL yes Acknowledge interrupt on exit yes Save IA32_PAT yes Load IA32_PAT yes Save IA32_EFER yes Load IA32_EFER yes Save VMX-preemption timer value yes VM-Entry controls Load debug controls default IA-32e mode guest yes Entry to SMM yes Deactivate dual-monitor treatment yes Load IA32_PERF_GLOBAL_CTRL yes Load IA32_PAT yes Load IA32_EFER yes Miscellaneous data VMX-preemption timer scale (log2) 5 Store EFER.LMA into IA-32e mode guest control no HLT activity state yes Shutdown activity state yes Wait-for-SIPI activity state yes IA32_SMBASE support no Number of CR3-target values 4 MSR-load/store count recommendation 0 IA32_SMM_MONITOR_CTL can be set to 1 no VMWRITE to VM-exit information fields no MSEG revision identifier 0 VPID and EPT capabilities Execute-only EPT translations yes Page-walk length 4 yes Paging-structure memory type UC yes Paging-structure memory type WB yes 2MB EPT pages yes 1GB EPT pages no INVEPT supported yes EPT accessed and dirty flags no Single-context INVEPT yes All-context INVEPT yes INVVPID supported yes Individual-address INVVPID yes Single-context INVVPID yes All-context INVVPID yes Single-context-retaining-globals INVVPID yes VM Functions EPTP Switching no
So for our test environment
rdtsc IS yielding VM exiting, and combined with
cpuid, giving huge performance penalty to Suricata
running inside a virtual machine.
We see that Docker incurs trivial resource overhead compared to bare metal, while KVM’s overhead is order of magnitude more than what Suricata itself uses. In terms of performance, while Docker imposes negligible, if any, penalty, KVM makes it much less efficient to run Suricata’s instructions and makes CPU a bottleneck at relatively low load levels.
The trace file
snort.log.1425823194 sends high volume of data in short period of time (about 22 seconds). Data is collected with granularity of 1 second.
Performance of Suricata
The final result of the tests is as follows.
|Bare Metal||1X||29 s||0||142202||157287081|
|Bare Metal||2X||28 s||0||284404||314574162|
|Bare Metal||4X||29 s||67126||500749||552477875|
Note: Time column is the amount of time it takes to reach stable value. Because we take medians of all samples to form a representation, at least half samples complete by that time.
We see that in terms of performance Docker setup is on a par with bare metal even when CPU is stressed at 4X load. Macvtap consumes CPU, resulting in DockerV setup dropping more packets than Docker setup at 4X load, but it is still in the same magnitude as Docker and bare metal setups.
The result of VM setup is noteworthy. Not only it drops significantly more packets, but a nontrivial portion of packets (358789 captured / 568808 total) isn’t even captured. In the following section, we will dive deeper into what happens to VM setup at 4X load.
Investigating the VM Issue
The following graph puts the cumulative number of packets sent (light blue line), cumulative number of packets received on the VM (orange line), cumulative number of packets captured (gray bar) / dropped (yellow bar) / decoded (dark blue bar) by Suricata, host CPU usage (green line, percentage), and VM memory usage (dark blue line, percentage) together:
Note that because we are taking medians, cumulative number of packets sent is not necessarily equal to cumulative number of packets received. I manually inspected several instances and for those instances they are indeed equal.
We see from the graph that Suricata packet capturing thread(s) never manages to process packets that come every single second. Most new packets are dropped and few are decoded. CPU usage reaches ceiling near second 10 and remains high for a few seconds. My guess is that because Suricata can’t read packets fast enough, the kernel buffer to hold packets becomes full and new packets are simply dropped. When Suricata catches up it no longer finds more packets to process.
After further checking, it turns out the the performance degradation is
largely due to the fact that X86
rdtsc instruction cannot
run efficiently in VM environment. It’s mostly used to get precise timestamps
for profiling purposes. Turning down rate of
profiling could mitigate the issue, but cannot solve it.
We see that container solutions like Docker are way more suitable than virtual machines for light-weight programs like Suricata in that (1) less resource is needed, and (2) the high cost of VM exiting triggered by some instructions can be saved.
Further: Experiment with Snort
We applied the same experiment to Snort. While the result of Snort is not as overwhelming as that we see on Suricata, performance penalty and resource overhead (VM setup eats 60% more CPU and 100% more RAM compared to Docker) are still significant. The data is also available on the GitHub repository.