-
Notifications
You must be signed in to change notification settings - Fork 1.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug] Processes get stuck after resuming VM from snapshot #4099
Comments
I'm not sure if this helps, but I randomly discovered that the issue can no longer be reproduced if I set the Also, I can't seem to reproduce this on an Intel machine (running host kernel v5.15). So maybe this issue is AMD-specific? I noticed my local AMD machine has a log line in the guest kernel showing Here are the full kernel boot logs: Logs[ 0.000000] Linux version 4.14.55-84.37.amzn2.x86_64 (mockbuild@ip-10-0-1-79) (gcc version 7.3.1 20180303 (Red Hat 7.3.1-5) (GCC)) #1 SMP Wed Jul 25 18:47:15 UTC 2018 [ 0.000000] Command line: ro console=ttyS0 reboot=k panic=1 pci=off nomodule random.trust_cpu=on i8042.noaux i8042.nomux i8042.nopnp i8042.dumbkbd tsc=reliable ipv6.disable=1 [ 0.000000] [Firmware Bug]: TSC doesn't count with P0 frequency! [ 0.000000] x86/fpu: Supporting XSAVE feature 0x001: 'x87 floating point registers' [ 0.000000] x86/fpu: Supporting XSAVE feature 0x002: 'SSE registers' [ 0.000000] x86/fpu: Supporting XSAVE feature 0x004: 'AVX registers' [ 0.000000] x86/fpu: Supporting XSAVE feature 0x020: 'AVX-512 opmask' [ 0.000000] x86/fpu: Supporting XSAVE feature 0x040: 'AVX-512 Hi256' [ 0.000000] x86/fpu: Supporting XSAVE feature 0x080: 'AVX-512 ZMM_Hi256' [ 0.000000] x86/fpu: Supporting XSAVE feature 0x200: 'Protection Keys User registers' [ 0.000000] x86/fpu: xstate_offset[2]: 576, xstate_sizes[2]: 256 [ 0.000000] x86/fpu: xstate_offset[5]: 832, xstate_sizes[5]: 64 [ 0.000000] x86/fpu: xstate_offset[6]: 896, xstate_sizes[6]: 512 [ 0.000000] x86/fpu: xstate_offset[7]: 1408, xstate_sizes[7]: 1024 [ 0.000000] x86/fpu: xstate_offset[9]: 2432, xstate_sizes[9]: 8 [ 0.000000] x86/fpu: Enabled xstate features 0x2e7, context size is 2440 bytes, using 'compacted' format. [ 0.000000] e820: BIOS-provided physical RAM map: [ 0.000000] BIOS-e820: [mem 0x0000000000000000-0x000000000009fbff] usable [ 0.000000] BIOS-e820: [mem 0x0000000000100000-0x0000000007ffffff] usable [ 0.000000] NX (Execute Disable) protection: active [ 0.000000] DMI not present or invalid. [ 0.000000] Hypervisor detected: KVM [ 0.000000] tsc: Fast TSC calibration using PIT [ 0.000000] e820: last_pfn = 0x8000 max_arch_pfn = 0x400000000 [ 0.000000] MTRR: Disabled [ 0.000000] x86/PAT: MTRRs disabled, skipping PAT initialization too. [ 0.000000] CPU MTRRs all blank - virtualized system. [ 0.000000] x86/PAT: Configuration [0-7]: WB WT UC- UC WB WT UC- UC [ 0.000000] found SMP MP-table at [mem 0x0009fc00-0x0009fc0f] mapped at [ffffffffff200c00] [ 0.000000] Scanning 1 areas for low memory corruption [ 0.000000] Using GB pages for direct mapping [ 0.000000] RAMDISK: [mem 0x07f24000-0x07ffffff] [ 0.000000] No NUMA configuration found [ 0.000000] Faking a node at [mem 0x0000000000000000-0x0000000007ffffff] [ 0.000000] NODE_DATA(0) allocated [mem 0x07f02000-0x07f23fff] [ 0.000000] kvm-clock: Using msrs 4b564d01 and 4b564d00 [ 0.000000] kvm-clock: cpu 0, msr 0:7f00001, primary cpu clock [ 0.000000] kvm-clock: using sched offset of 89597168 cycles [ 0.000000] clocksource: kvm-clock: mask: 0xffffffffffffffff max_cycles: 0x1cd42e4dffb, max_idle_ns: 881590591483 ns [ 0.000000] Zone ranges: [ 0.000000] DMA [mem 0x0000000000001000-0x0000000000ffffff] [ 0.000000] DMA32 [mem 0x0000000001000000-0x0000000007ffffff] [ 0.000000] Normal empty [ 0.000000] Movable zone start for each node [ 0.000000] Early memory node ranges [ 0.000000] node 0: [mem 0x0000000000001000-0x000000000009efff] [ 0.000000] node 0: [mem 0x0000000000100000-0x0000000007ffffff] [ 0.000000] Initmem setup node 0 [mem 0x0000000000001000-0x0000000007ffffff] [ 0.000000] Intel MultiProcessor Specification v1.4 [ 0.000000] MPTABLE: OEM ID: FC [ 0.000000] MPTABLE: Product ID: 000000000000 [ 0.000000] MPTABLE: APIC at: 0xFEE00000 [ 0.000000] Processor #0 (Bootup-CPU) [ 0.000000] IOAPIC[0]: apic_id 2, version 17, address 0xfec00000, GSI 0-23 [ 0.000000] Processors: 1 [ 0.000000] smpboot: Allowing 1 CPUs, 0 hotplug CPUs [ 0.000000] PM: Registered nosave memory: [mem 0x00000000-0x00000fff] [ 0.000000] PM: Registered nosave memory: [mem 0x0009f000-0x000fffff] [ 0.000000] e820: [mem 0x08000000-0xffffffff] available for PCI devices [ 0.000000] Booting paravirtualized kernel on KVM [ 0.000000] clocksource: refined-jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 7645519600211568 ns [ 0.000000] random: get_random_bytes called from start_kernel+0x94/0x486 with crng_init=0 [ 0.000000] setup_percpu: NR_CPUS:128 nr_cpumask_bits:128 nr_cpu_ids:1 nr_node_ids:1 [ 0.000000] percpu: Embedded 41 pages/cpu @ffff880007c00000 s128728 r8192 d31016 u2097152 [ 0.000000] KVM setup async PF for cpu 0 [ 0.000000] kvm-stealtime: cpu 0, msr 7c15040 [ 0.000000] PV qspinlock hash table entries: 256 (order: 0, 4096 bytes) [ 0.000000] Built 1 zonelists, mobility grouping on. Total pages: 32137 [ 0.000000] Policy zone: DMA32 [ 0.000000] Kernel command line: ro console=ttyS0 reboot=k panic=1 pci=off nomodule random.trust_cpu=on i8042.noaux i8042.nomux i8042.nopnp i8042.dumbkbd tsc=reliable ipv6.disable=1 [ 0.000000] PID hash table entries: 512 (order: 0, 4096 bytes) [ 0.000000] Memory: 110184K/130680K available (8204K kernel code, 622K rwdata, 1464K rodata, 1268K init, 2820K bss, 20496K reserved, 0K cma-reserved) [ 0.000000] SLUB: HWalign=64, Order=0-3, MinObjects=0, CPUs=1, Nodes=1 [ 0.004000] Hierarchical RCU implementation. [ 0.004000] RCU restricting CPUs from NR_CPUS=128 to nr_cpu_ids=1. [ 0.004000] RCU: Adjusting geometry for rcu_fanout_leaf=16, nr_cpu_ids=1 [ 0.004000] NR_IRQS: 4352, nr_irqs: 48, preallocated irqs: 16 [ 0.004000] Console: colour dummy device 80x25 [ 0.004000] console [ttyS0] enabled [ 0.004000] tsc: Detected 4491.540 MHz processor [ 0.004000] Calibrating delay loop (skipped) preset value.. 8983.08 BogoMIPS (lpj=17966160) [ 0.004000] pid_max: default: 32768 minimum: 301 [ 0.004000] Security Framework initialized [ 0.004000] SELinux: Initializing. [ 0.004000] Dentry cache hash table entries: 16384 (order: 5, 131072 bytes) [ 0.004000] Inode-cache hash table entries: 8192 (order: 4, 65536 bytes) [ 0.004000] Mount-cache hash table entries: 512 (order: 0, 4096 bytes) [ 0.004000] Mountpoint-cache hash table entries: 512 (order: 0, 4096 bytes) [ 0.004000] Last level iTLB entries: 4KB 512, 2MB 512, 4MB 256 [ 0.004000] Last level dTLB entries: 4KB 3072, 2MB 3072, 4MB 1536, 1GB 0 [ 0.004000] Spectre V2 : Mitigation: Full AMD retpoline [ 0.004000] Spectre V2 : Spectre v2 mitigation: Enabling Indirect Branch Prediction Barrier [ 0.004000] Spectre V2 : Enabling Restricted Speculation for firmware calls [ 0.004000] Speculative Store Bypass: Mitigation: Speculative Store Bypass disabled via prctl and seccomp [ 0.006983] Freeing SMP alternatives memory: 28K [ 0.007832] smpboot: Max logical packages: 1 [ 0.008055] x2apic enabled [ 0.008213] Switched APIC routing to physical x2apic. [ 0.008780] ..TIMER: vector=0x30 apic1=0 pin1=0 apic2=-1 pin2=-1 [ 0.008980] smpboot: CPU0: AMD EPYC (family: 0x19, model: 0x61, stepping: 0x2) [ 0.009245] Performance Events: [ 0.009247] core perfctr but no constraints; unknown hardware! [ 0.009528] no PMU driver, software events only. [ 0.009711] Hierarchical SRCU implementation. [ 0.011485] smp: Bringing up secondary CPUs ... [ 0.011626] smp: Brought up 1 node, 1 CPU [ 0.011752] smpboot: Total of 1 processors activated (8983.08 BogoMIPS) [ 0.012000] devtmpfs: initialized [ 0.012000] x86/mm: Memory block size: 128MB [ 0.012000] clocksource: jiffies: mask: 0xffffffff max_cycles: 0xffffffff, max_idle_ns: 7645041785100000 ns [ 0.012000] futex hash table entries: 256 (order: 2, 16384 bytes) [ 0.012152] NET: Registered protocol family 16 [ 0.012359] cpuidle: using governor ladder [ 0.012488] cpuidle: using governor menu [ 0.014151] HugeTLB registered 1.00 GiB page size, pre-allocated 0 pages [ 0.014359] HugeTLB registered 2.00 MiB page size, pre-allocated 0 pages [ 0.014640] dmi: Firmware registration failed. [ 0.014818] NetLabel: Initializing [ 0.014924] NetLabel: domain hash size = 128 [ 0.015059] NetLabel: protocols = UNLABELED CIPSOv4 CALIPSO [ 0.015241] NetLabel: unlabeled traffic allowed by default [ 0.015444] clocksource: Switched to clocksource kvm-clock [ 0.015624] VFS: Disk quotas dquot_6.6.0 [ 0.015748] VFS: Dquot-cache hash table entries: 512 (order 0, 4096 bytes) [ 0.016000] NET: Registered protocol family 2 [ 0.016000] TCP established hash table entries: 1024 (order: 1, 8192 bytes) [ 0.016000] TCP bind hash table entries: 1024 (order: 2, 16384 bytes) [ 0.016000] TCP: Hash tables configured (established 1024 bind 1024) [ 0.016000] UDP hash table entries: 256 (order: 1, 8192 bytes) [ 0.016000] UDP-Lite hash table entries: 256 (order: 1, 8192 bytes) [ 0.016000] NET: Registered protocol family 1 [ 0.016000] Unpacking initramfs... [ 0.016000] Freeing initrd memory: 880K [ 0.016000] platform rtc_cmos: registered platform RTC device (no PNP device found) [ 0.016000] Scanning for low memory corruption every 60 seconds [ 0.016350] audit: initializing netlink subsys (disabled) [ 0.016640] Initialise system trusted keyrings [ 0.016780] Key type blacklist registered [ 0.016921] audit: type=2000 audit(1694379250.871:1): state=initialized audit_enabled=0 res=1 [ 0.017192] workingset: timestamp_bits=36 max_order=15 bucket_order=0 [ 0.017793] squashfs: version 4.0 (2009/01/31) Phillip Lougher [ 0.018927] Key type asymmetric registered [ 0.019056] Asymmetric key parser 'x509' registered [ 0.019214] Block layer SCSI generic (bsg) driver version 0.4 loaded (major 254) [ 0.019452] io scheduler noop registered (default) [ 0.019617] io scheduler cfq registered [ 0.019764] Serial: 8250/16550 driver, 1 ports, IRQ sharing disabled [ 0.040269] serial8250: ttyS0 at I/O 0x3f8 (irq = 4, base_baud = 115200) is a U6_16550A [ 0.041195] loop: module loaded [ 0.041300] tun: Universal TUN/TAP device driver, 1.6 [ 0.041472] hidraw: raw HID events driver (C) Jiri Kosina [ 0.041657] nf_conntrack version 0.5.0 (1024 buckets, 4096 max) [ 0.041893] ip_tables: (C) 2000-2006 Netfilter Core Team [ 0.042070] Initializing XFRM netlink socket [ 0.042206] IPv6: Loaded, but administratively disabled, reboot required to enable [ 0.042438] NET: Registered protocol family 17 [ 0.042582] Bridge firewalling registered [ 0.042810] sched_clock: Marking stable (40263980, 0)->(65107117, -24843137) [ 0.043068] registered taskstats version 1 [ 0.043197] Loading compiled-in X.509 certificates [ 0.043628] Loaded X.509 cert 'Build time autogenerated kernel key: 3472798b31ba23b86c1c5c7236c9c91723ae5ee9' [ 0.043943] zswap: default zpool zbud not available [ 0.044111] zswap: pool creation failed [ 0.044354] Key type encrypted registered [ 0.044957] Freeing unused kernel memory: 1268K [ 0.052047] Write protecting the kernel read-only data: 12288k [ 0.053097] Freeing unused kernel memory: 2016K [ 0.053771] Freeing unused kernel memory: 584K |
Thanks for reporting this. At the moment, firecracker officially supports host kernel 4.14 and 5.10 [README.md] and will add host kernel 6.1 support. Are you able to reproduce it on host kernel 5.10 and 6.1, although we'll also try to reproduce the issue on our side? |
I don't have easy access to an AMD machine with any of those kernel versions, but later this week I will try and see if I can get 5.10 or 6.1 working with my current setup. Does Firecracker have any test suites that run against AMD machines? (I suspect this may be AMD-specific.) I would be happy to contribute a test that demonstrates the issue - that would be a lot easier for me than installing a different kernel version, and hopefully easier to find a fix as well once there's a failing test. |
Yes. m6a.metal is AMD instance that firecracker is tested on. Thanks for your help and effort! |
Ah OK, I will actually just spin one of those up on AWS then :) Thanks |
I was able to reproduce this on an m6a.metal instance with Linux 5.10. I did Once SSH'd,
If it helps, I had to run |
Here is what it looks like when running the repro on that EC2 instance - whenever it prints simplescreenrecorder-2023-09-11_10.20.48.mp4The kernel logs on this m6a.metal instance show the same error message in the kernel logs:
|
I've tried codesandbox's fork of firecracker which includes this commit: codesandbox@4164371 I am not sure if this is the "right" fix, but the issue no longer reproduces when using this commit. Any thoughts on whether this commit should be upstreamed? |
nanosleep
get stuck after resuming VM from snapshot
Not sure this is related, but our implementation of container snapshots backed by firecracker VMs using firecracker-containerd in vHive stumbles upon a similar problem: it seems like firecracker-containerd's in-VM agent stops responding and hangs after a VM snapshot is created and the VM is resumed (vhive-serverless/vHive#818). We use a default setup suggested by firecracker-containerd. |
@CuriousGeorgiy are you hitting this issue on an AMD CPU, or Intel? Also, could you try running the repro steps I posted in the issue description, and letting us know if you see the same problem that I described? This would be very helpful as an extra data point 🙏 |
@bduffany I am renting a node on CloudLab with an Intel CPU. I tried running the repro steps you posted, but after waiting for 100 rounds I didn't see any problems you described. |
@CuriousGeorgiy thanks for running the repro steps! It is interesting that you couldn't repro the issue on an Intel machine. It seems increasingly likely that the original issue I reported is AMD-specific, since I couldn't reproduce it on Intel either. I think the issue that you are seeing is probably different than the one reported here, since the repro script doesn't exhibit the problematic behavior when you tried it. It may be worth reporting a new issue to the firecracker-containerd repo, with more details about your setup/environment and ideally a minimal reproducer. |
Hi @bduffany, thanks for providing the repro steps. We are working on this issue and to provide an update :
|
Friendly ping, is this issue still on your radar? This is still a frequent pain point for us when using Firecracker on AMD machines, and has us effectively locked in to Intel which is not very desirable. |
Hi @bduffany, we are tracking this as one of our high priority task. We'll let you know as soon as we have an update on this. |
Hi @bduffany. We are still working on the root cause of this issue. We will update you as soon as possible. |
Hi @bduffany, I am currently looking at this. I found that passing |
@pb8o thank you for looking into this! I tried setting |
hey @pb8o, curious whether there are any leads on this that you can share? We would love to be able to run our Firecracker workloads on AMD machines, because they are much cheaper. Happy to help any way that I can - not much of an expert on virtualization but can maybe help try to find a more consistent repro, if that is part of what makes this tricky? (And thank you again for looking into this!) |
Hi @bduffany, Sorry for the delay, unfortunately we were not able to make progress regarding this issue. |
FYI, we started hitting this again, but on Intel this time (on GCP under nested virtualization), after GCP upgraded our nodes from kernel version We never quite figured out how to map those GCP kernel versions to upstream kernel versions, or how to figure out which commits are included between those versions. Maybe someone with contacts at Google Cloud could help 😅 |
AMD hosts do not have TSC deadline enabled, and that seems to cause issues like in firecracker-microvm#4099. Only expose TSC_DEADLINE to the guest if the host supports it. Closes firecracker-microvm#4099 Signed-off-by: Pablo Barbáchano <[email protected]>
AMD hosts do not have TSC deadline enabled, and that seems to cause issues like in firecracker-microvm#4099. Only expose TSC_DEADLINE to the guest if the host supports it. Closes firecracker-microvm#4099 Signed-off-by: Pablo Barbáchano <[email protected]>
AMD hosts do not have TSC deadline enabled, and that seems to cause issues like in firecracker-microvm#4099. Only expose TSC_DEADLINE to the guest if the host supports it. Closes firecracker-microvm#4099 Signed-off-by: Pablo Barbáchano <[email protected]>
AMD hosts do not have TSC deadline enabled, and that seems to cause issues like in firecracker-microvm#4099. Only expose TSC_DEADLINE to the guest if the host supports it. Closes firecracker-microvm#4099 Signed-off-by: Pablo Barbáchano <[email protected]>
Hi @bduffany, first of all thanks for your patience. We are sorry that it is taking so long to root cause and the time between updates, and I wanted to reassure you that we take this issue seriously and we are still investigating a possible solution. Regarding the patch
I was able to reproduce the issue on Ubuntu 24.04, which had a 6.8.0 kernel with that patch. I was also able to reproduce the issue with newer guest kernels and newer AMD CPUs like Genoa, so we can rule out something that is fixed in newer kernels or CPUs. At this point we believe the issue is that we use TSC deadline feature in both Intel and AMD. Intel does support TSC deadline, so works fine. In the AMD CPUs we tested however, TSC deadline is not enabled but Firecracker still presents it to the guest via We have tried working around the issue in #4655 by only enabling TSC deadline if it's enabled on the host. This works fine, but creates a boot time regression of 40-50% in our tests. We are currently trying to understand where this time is going and if we can mitigate it. |
AMD hosts do not have TSC deadline enabled, and that seems to cause issues like in firecracker-microvm#4099. Only expose TSC_DEADLINE to the guest if the host supports it. Closes firecracker-microvm#4099 Signed-off-by: Pablo Barbáchano <[email protected]>
The `wait_for_up()` sends a dummy SSH command with a timeout of 10s, with the intention of blocking the integration test until a VM is responsive after booting. However, post-snapshot restore it runs into issue firecracker-microvm#4099, where on AMD inside the guest, sshd gets stuck inside a nanosleep syscall for longer than intended (and thus longer than the 10s timeout). Alternatively, we could pass `lapic=notscdeadline` in the kernel commandline arguments, but we do not have an easy way to _append_ to the default kernel command line from integration tests, and I'd like to avoid duplicating our default across a bunch of places. Signed-off-by: Patrick Roy <[email protected]>
The `wait_for_up()` sends a dummy SSH command with a timeout of 10s, with the intention of blocking the integration test until a VM is responsive after booting. However, post-snapshot restore it runs into issue #4099, where on AMD inside the guest, sshd gets stuck inside a nanosleep syscall for longer than intended (and thus longer than the 10s timeout). Alternatively, we could pass `lapic=notscdeadline` in the kernel commandline arguments, but we do not have an easy way to _append_ to the default kernel command line from integration tests, and I'd like to avoid duplicating our default across a bunch of places. Signed-off-by: Patrick Roy <[email protected]>
When restoring MSRs, we currently do not enforce any sort of relative ordering. However, some MSRs have implicit dependencies on other MSRs being restored before them, and failing to fulfill these dependencies can result in incorrect VM behavior after resuming. One example of such an implicit dependency between MSRs is the pair (`MSR_IA32_TSC_DEADLINE`, `MSR_IA32_TSC`). When restoring `MSR_IA32_TSC_DEADLINE`, KVM internally checks whether the value of this restoration implies that the guest was waiting for the tsc_deadline timer to expire at the time of being paused for snapshotting. If yes, it primes a (either harddware or software depending on support) timer on the host to make sure the guest will receive the expected interrupt after restoration. To determine whether this is needed, KVM looks at the guest's timestamp counter (TSC) and compares it with the requested tsc_deadline value. The value KVM reads for the guest's TSC depends on the value of MSR_IA32_TSC. Thus, if MSR_IA32_TSC_DEADLINE is set before MSR_IA32_TSC is restored, this comparison will yield a wrong result (as the deadline value is compared with something uninitialized). This can either result in KVM determining the guest wasn't waiting for a timing expiry at the time of snapshotting, or cause it to schedule the timer interrupt too far into the future. Note that the former is only a problem on AMD platforms, which do not support the TSC_DEADLINE feature at the hardware level. Here, KVM falls back to a software timer, which explicitly does not notify the vCPU if the deadline value is "in the past". The hardware timer used on other x86 platforms on the other hand always fires (potentially firing immediately if the deadline value is in the past). This commit fixes the above by ensuring we always restore MSR_IA32_TSC before restoring MSR_IA32_TSC_DEADLINE. Fixes firecracker-microvm#4099 Signed-off-by: Patrick Roy <[email protected]>
When restoring MSRs, we currently do not enforce any sort of relative ordering. However, some MSRs have implicit dependencies on other MSRs being restored before them, and failing to fulfill these dependencies can result in incorrect VM behavior after resuming. One example of such an implicit dependency between MSRs is the pair (`MSR_IA32_TSC_DEADLINE`, `MSR_IA32_TSC`). When restoring `MSR_IA32_TSC_DEADLINE`, KVM internally checks whether the value of this restoration implies that the guest was waiting for the tsc_deadline timer to expire at the time of being paused for snapshotting. If yes, it primes a (either harddware or software depending on support) timer on the host to make sure the guest will receive the expected interrupt after restoration. To determine whether this is needed, KVM looks at the guest's timestamp counter (TSC) and compares it with the requested tsc_deadline value. The value KVM reads for the guest's TSC depends on the value of MSR_IA32_TSC. Thus, if MSR_IA32_TSC_DEADLINE is set before MSR_IA32_TSC is restored, this comparison will yield a wrong result (as the deadline value is compared with something uninitialized). This can either result in KVM determining the guest wasn't waiting for a timing expiry at the time of snapshotting, or cause it to schedule the timer interrupt too far into the future. Note that the former is only a problem on AMD platforms, which do not support the TSC_DEADLINE feature at the hardware level. Here, KVM falls back to a software timer, which explicitly does not notify the vCPU if the deadline value is "in the past". The hardware timer used on other x86 platforms on the other hand always fires (potentially firing immediately if the deadline value is in the past). This commit fixes the above by ensuring we always restore MSR_IA32_TSC before restoring MSR_IA32_TSC_DEADLINE. We realize this by splitting the lists of MSRs that KVM gives us into one additional chunk containing all "deferred" MSRs that needs to be restored "as late as possible". This splitting happens at snapshot creation time, to get it off the hot-path. Fixes firecracker-microvm#4099 Signed-off-by: Patrick Roy <[email protected]>
When restoring MSRs, we currently do not enforce any sort of relative ordering. However, some MSRs have implicit dependencies on other MSRs being restored before them, and failing to fulfill these dependencies can result in incorrect VM behavior after resuming. One example of such an implicit dependency between MSRs is the pair (`MSR_IA32_TSC_DEADLINE`, `MSR_IA32_TSC`). When restoring `MSR_IA32_TSC_DEADLINE`, KVM internally checks whether the value of this restoration implies that the guest was waiting for the tsc_deadline timer to expire at the time of being paused for snapshotting. If yes, it primes a (either harddware or software depending on support) timer on the host to make sure the guest will receive the expected interrupt after restoration. To determine whether this is needed, KVM looks at the guest's timestamp counter (TSC) and compares it with the requested tsc_deadline value. The value KVM reads for the guest's TSC depends on the value of MSR_IA32_TSC. Thus, if MSR_IA32_TSC_DEADLINE is set before MSR_IA32_TSC is restored, this comparison will yield a wrong result (as the deadline value is compared with something uninitialized). This can either result in KVM determining the guest wasn't waiting for a timing expiry at the time of snapshotting, or cause it to schedule the timer interrupt too far into the future. Note that the former is only a problem on AMD platforms, which do not support the TSC_DEADLINE feature at the hardware level. Here, KVM falls back to a software timer, which explicitly does not notify the vCPU if the deadline value is "in the past". The hardware timer used on other x86 platforms on the other hand always fires (potentially firing immediately if the deadline value is in the past). This commit fixes the above by ensuring we always restore MSR_IA32_TSC before restoring MSR_IA32_TSC_DEADLINE. We realize this by splitting the lists of MSRs that KVM gives us into one additional chunk containing all "deferred" MSRs that needs to be restored "as late as possible". This splitting happens at snapshot creation time, to get it off the hot-path. Fixes firecracker-microvm#4099 Signed-off-by: Patrick Roy <[email protected]>
add changelog entry about firecracker-microvm#4099 Signed-off-by: Patrick Roy <[email protected]>
Good morning @bduffany, With the commit from #4666, I am no longer able to reproduce the issue. Could you also give this a try? Thanks! To follow up on some loose ends from @pb8o's last update: The alternate fix of simply disabling the TSC_DEADLINE feature (either via CPUID or |
When restoring MSRs, we currently do not enforce any sort of relative ordering. However, some MSRs have implicit dependencies on other MSRs being restored before them, and failing to fulfill these dependencies can result in incorrect VM behavior after resuming. One example of such an implicit dependency between MSRs is the pair (`MSR_IA32_TSC_DEADLINE`, `MSR_IA32_TSC`). When restoring `MSR_IA32_TSC_DEADLINE`, KVM internally checks whether the value of this restoration implies that the guest was waiting for the tsc_deadline timer to expire at the time of being paused for snapshotting. If yes, it primes a (either harddware or software depending on support) timer on the host to make sure the guest will receive the expected interrupt after restoration. To determine whether this is needed, KVM looks at the guest's timestamp counter (TSC) and compares it with the requested tsc_deadline value. The value KVM reads for the guest's TSC depends on the value of MSR_IA32_TSC. Thus, if MSR_IA32_TSC_DEADLINE is set before MSR_IA32_TSC is restored, this comparison will yield a wrong result (as the deadline value is compared with something uninitialized). This can either result in KVM determining the guest wasn't waiting for a timing expiry at the time of snapshotting, or cause it to schedule the timer interrupt too far into the future. Note that the former is only a problem on AMD platforms, which do not support the TSC_DEADLINE feature at the hardware level. Here, KVM falls back to a software timer, which explicitly does not notify the vCPU if the deadline value is "in the past". The hardware timer used on other x86 platforms on the other hand always fires (potentially firing immediately if the deadline value is in the past). This commit fixes the above by ensuring we always restore MSR_IA32_TSC before restoring MSR_IA32_TSC_DEADLINE. We realize this by splitting the lists of MSRs that KVM gives us into one additional chunk containing all "deferred" MSRs that needs to be restored "as late as possible". This splitting happens at snapshot creation time, to get it off the hot-path. Fixes firecracker-microvm#4099 Signed-off-by: Patrick Roy <[email protected]>
add changelog entry about firecracker-microvm#4099 Signed-off-by: Patrick Roy <[email protected]>
add changelog entry about #4099 Signed-off-by: Patrick Roy <[email protected]>
When restoring MSRs, we currently do not enforce any sort of relative ordering. However, some MSRs have implicit dependencies on other MSRs being restored before them, and failing to fulfill these dependencies can result in incorrect VM behavior after resuming. One example of such an implicit dependency between MSRs is the pair (`MSR_IA32_TSC_DEADLINE`, `MSR_IA32_TSC`). When restoring `MSR_IA32_TSC_DEADLINE`, KVM internally checks whether the value of this restoration implies that the guest was waiting for the tsc_deadline timer to expire at the time of being paused for snapshotting. If yes, it primes a (either harddware or software depending on support) timer on the host to make sure the guest will receive the expected interrupt after restoration. To determine whether this is needed, KVM looks at the guest's timestamp counter (TSC) and compares it with the requested tsc_deadline value. The value KVM reads for the guest's TSC depends on the value of MSR_IA32_TSC. Thus, if MSR_IA32_TSC_DEADLINE is set before MSR_IA32_TSC is restored, this comparison will yield a wrong result (as the deadline value is compared with something uninitialized). This can either result in KVM determining the guest wasn't waiting for a timing expiry at the time of snapshotting, or cause it to schedule the timer interrupt too far into the future. Note that the former is only a problem on AMD platforms, which do not support the TSC_DEADLINE feature at the hardware level. Here, KVM falls back to a software timer, which explicitly does not notify the vCPU if the deadline value is "in the past". The hardware timer used on other x86 platforms on the other hand always fires (potentially firing immediately if the deadline value is in the past). This commit fixes the above by ensuring we always restore MSR_IA32_TSC before restoring MSR_IA32_TSC_DEADLINE. We realize this by splitting the lists of MSRs that KVM gives us into one additional chunk containing all "deferred" MSRs that needs to be restored "as late as possible". This splitting happens at snapshot creation time, to get it off the hot-path. Fixes firecracker-microvm#4099 Signed-off-by: Patrick Roy <[email protected]>
add changelog entry about firecracker-microvm#4099 Signed-off-by: Patrick Roy <[email protected]>
When restoring MSRs, we currently do not enforce any sort of relative ordering. However, some MSRs have implicit dependencies on other MSRs being restored before them, and failing to fulfill these dependencies can result in incorrect VM behavior after resuming. One example of such an implicit dependency between MSRs is the pair (`MSR_IA32_TSC_DEADLINE`, `MSR_IA32_TSC`). When restoring `MSR_IA32_TSC_DEADLINE`, KVM internally checks whether the value of this restoration implies that the guest was waiting for the tsc_deadline timer to expire at the time of being paused for snapshotting. If yes, it primes a (either harddware or software depending on support) timer on the host to make sure the guest will receive the expected interrupt after restoration. To determine whether this is needed, KVM looks at the guest's timestamp counter (TSC) and compares it with the requested tsc_deadline value. The value KVM reads for the guest's TSC depends on the value of MSR_IA32_TSC. Thus, if MSR_IA32_TSC_DEADLINE is set before MSR_IA32_TSC is restored, this comparison will yield a wrong result (as the deadline value is compared with something uninitialized). This can either result in KVM determining the guest wasn't waiting for a timing expiry at the time of snapshotting, or cause it to schedule the timer interrupt too far into the future. Note that the former is only a problem on AMD platforms, which do not support the TSC_DEADLINE feature at the hardware level. Here, KVM falls back to a software timer, which explicitly does not notify the vCPU if the deadline value is "in the past". The hardware timer used on other x86 platforms on the other hand always fires (potentially firing immediately if the deadline value is in the past). This commit fixes the above by ensuring we always restore MSR_IA32_TSC before restoring MSR_IA32_TSC_DEADLINE. We realize this by splitting the lists of MSRs that KVM gives us into one additional chunk containing all "deferred" MSRs that needs to be restored "as late as possible". This splitting happens at snapshot creation time, to get it off the hot-path. Fixes #4099 Signed-off-by: Patrick Roy <[email protected]>
add changelog entry about #4099 Signed-off-by: Patrick Roy <[email protected]>
@roypat Thank you so much for the fix! I verified on my AMD machine that the script is able to complete 100 iterations without getting stuck, without the |
The `wait_for_up()` sends a dummy SSH command with a timeout of 10s, with the intention of blocking the integration test until a VM is responsive after booting. However, post-snapshot restore it runs into issue firecracker-microvm#4099, where on AMD inside the guest, sshd gets stuck inside a nanosleep syscall for longer than intended (and thus longer than the 10s timeout). Alternatively, we could pass `lapic=notscdeadline` in the kernel commandline arguments, but we do not have an easy way to _append_ to the default kernel command line from integration tests, and I'd like to avoid duplicating our default across a bunch of places. Signed-off-by: Patrick Roy <[email protected]>
When restoring MSRs, we currently do not enforce any sort of relative ordering. However, some MSRs have implicit dependencies on other MSRs being restored before them, and failing to fulfill these dependencies can result in incorrect VM behavior after resuming. One example of such an implicit dependency between MSRs is the pair (`MSR_IA32_TSC_DEADLINE`, `MSR_IA32_TSC`). When restoring `MSR_IA32_TSC_DEADLINE`, KVM internally checks whether the value of this restoration implies that the guest was waiting for the tsc_deadline timer to expire at the time of being paused for snapshotting. If yes, it primes a (either harddware or software depending on support) timer on the host to make sure the guest will receive the expected interrupt after restoration. To determine whether this is needed, KVM looks at the guest's timestamp counter (TSC) and compares it with the requested tsc_deadline value. The value KVM reads for the guest's TSC depends on the value of MSR_IA32_TSC. Thus, if MSR_IA32_TSC_DEADLINE is set before MSR_IA32_TSC is restored, this comparison will yield a wrong result (as the deadline value is compared with something uninitialized). This can either result in KVM determining the guest wasn't waiting for a timing expiry at the time of snapshotting, or cause it to schedule the timer interrupt too far into the future. Note that the former is only a problem on AMD platforms, which do not support the TSC_DEADLINE feature at the hardware level. Here, KVM falls back to a software timer, which explicitly does not notify the vCPU if the deadline value is "in the past". The hardware timer used on other x86 platforms on the other hand always fires (potentially firing immediately if the deadline value is in the past). This commit fixes the above by ensuring we always restore MSR_IA32_TSC before restoring MSR_IA32_TSC_DEADLINE. We realize this by splitting the lists of MSRs that KVM gives us into one additional chunk containing all "deferred" MSRs that needs to be restored "as late as possible". This splitting happens at snapshot creation time, to get it off the hot-path. Fixes firecracker-microvm#4099 Signed-off-by: Patrick Roy <[email protected]>
add changelog entry about firecracker-microvm#4099 Signed-off-by: Patrick Roy <[email protected]>
Describe the bug
After resuming a VM from snapshot, processes occasionally get stuck. A minimal example is an
init
binary that just runswhile (true) { sleep 100ms ; print 'hello' }
- after resuming from snapshot, it only sometimes is able to resume the loop, and other times it gets stuck and does not print anything at all.To Reproduce
make
,cc
(C compiler, fromgcc
package),glibc-static
(for statically linking the init binary), andjq
(for parsing firecracker API output)make test
make test
is doing the following:tail -f
in the background)Expected behaviour
When running
make test
in that repo, the init binary should printrunning
several times after each resume. But on some resumes, it appears stuck, and does not print anything until the next resume.Interesting details:
nanosleep
syscalls with an NOP loop of 1e9 iterations (for (int i=0; i<1e9; i++) continue;
)firecracker
binary in between, but could not reproduce.Environment
m6a.metal
instancem6a.metal
instance)Additional context
This repro above is a minimal example of a much more troublesome issue where we are having trouble reconnecting to microVMs after resuming them from snapshots. We are running a server inside the VM and are having trouble connecting to over vsock. I suspected that the guest process was "stuck" somehow since when running
sleep 1 && print('hello')
in a background loop, it sometimes doesn't print anything. I came up with this minimal reproducer for this behavior.Checks
The text was updated successfully, but these errors were encountered: