virt_mshv_vtl/processor/hardware_cvm/
tlb_lock.rs

1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
// Copyright (c) Microsoft Corporation.
// Licensed under the MIT License.

//! TLB lock infrastructure support for hardware-isolated partitions.

use crate::UhCvmPartitionState;
use crate::UhProcessor;
use crate::processor::HardwareIsolatedBacking;
use hcl::GuestVtl;
use hvdef::Vtl;
use std::sync::atomic::Ordering;
use virt::VpIndex;

pub(crate) struct TlbLockAccess<'a> {
    pub vp_index: VpIndex,
    pub cvm_partition: &'a UhCvmPartitionState,
}

impl TlbLockAccess<'_> {
    pub fn set_wait_for_tlb_locks(&mut self, target_vtl: GuestVtl) {
        // Capture the set of VPs that are currently holding the TLB lock. Only
        // those VPs that hold the lock at this point can block progress, because
        // any VP that acquires the lock after this point is guaranteed to see
        // state that this VP has already flushed.
        let self_index = self.vp_index.index() as usize;
        let self_lock = &self.cvm_partition.vp_inner(self_index as u32).tlb_lock_info[target_vtl];
        for vp in self.cvm_partition.tlb_locked_vps[target_vtl]
            .clone()
            .iter_ones()
        {
            // Never wait on the current VP, since the current VP will always
            // release its locks correctly when returning to the target VTL.
            if vp == self_index {
                continue;
            }

            // First record that this VP will be waiting on the target VP.
            // Because the lock may already have been requested by a lower VTL,
            // the current VP may already be waiting for this target, and if so,
            // it should not count as an additional blocking VP.
            if self_lock.blocking_vps.set_aliased(vp, true) {
                continue;
            }
            self_lock.blocking_vp_count.fetch_add(1, Ordering::Relaxed);

            // Now advise the target VP that it is blocking this VP.
            // Because the wait by the current VP on the target VP is known to
            // be new, this bit should not already be set.
            let other_lock_blocked =
                &self.cvm_partition.vp_inner(vp as u32).tlb_lock_info[target_vtl].blocked_vps;
            let _was_other_lock_blocked = other_lock_blocked.set_aliased(self_index, true);
            debug_assert!(!_was_other_lock_blocked);

            // It is possible that the target VP released the TLB lock before
            // the current VP was added to its blocked set. Check again to
            // see whether the TLB lock is still held, and if not, remove the
            // block.
            if !self.cvm_partition.tlb_locked_vps[target_vtl][vp] {
                other_lock_blocked.set_aliased(self_index, false);
                if self_lock.blocking_vps.set_aliased(vp, false) {
                    self_lock.blocking_vp_count.fetch_sub(1, Ordering::Relaxed);
                }
            }
        }
    }
}

#[expect(private_bounds)]
impl<B: HardwareIsolatedBacking> UhProcessor<'_, B> {
    /// Causes the specified VTL on the current VP to wait on all TLB locks.
    /// This is typically used to synchronize VTL permission changes with
    /// concurrent instruction emulation.
    pub(crate) fn set_wait_for_tlb_locks(&mut self, target_vtl: GuestVtl) {
        TlbLockAccess {
            vp_index: self.vp_index(),
            cvm_partition: B::cvm_partition_state(self.shared),
        }
        .set_wait_for_tlb_locks(target_vtl);
    }

    /// Lock the TLB of the target VTL on the current VP.
    pub(crate) fn set_tlb_lock(&mut self, requesting_vtl: Vtl, target_vtl: GuestVtl) {
        debug_assert!(requesting_vtl > Vtl::from(target_vtl));

        self.cvm_partition().tlb_locked_vps[target_vtl]
            .set_aliased(self.vp_index().index() as usize, true);
        self.vtls_tlb_locked.set(requesting_vtl, target_vtl, true);
    }

    /// Unlocks the TLBs of a specific lower VTL
    pub(crate) fn unlock_tlb_lock_target(&mut self, unlocking_vtl: Vtl, target_vtl: GuestVtl) {
        debug_assert!(unlocking_vtl != Vtl::Vtl0);

        let self_index = self.vp_index().index() as usize;
        // If this VP hasn't taken a lock, no need to do anything.
        if self.vtls_tlb_locked.get(unlocking_vtl, target_vtl) {
            self.vtls_tlb_locked.set(unlocking_vtl, target_vtl, false);
            // A memory fence is required after indicating that the target VTL is no
            // longer locked, because other VPs will make decisions about how to
            // handle blocking based on this information, and the loop below relies on
            // those processors having an accurate view of the lock state.
            std::sync::atomic::fence(Ordering::SeqCst);

            // If the lock for VTL 0 is being released by VTL 2, then check
            // to see whether VTL 1 also holds a lock for VTL 0. If so, no
            // wait can be unblocked until VTL 1 also releases its lock.
            if unlocking_vtl == Vtl::Vtl2
                && target_vtl == GuestVtl::Vtl0
                && self.vtls_tlb_locked.get(Vtl::Vtl1, GuestVtl::Vtl0)
            {
                return;
            }

            // Now we can remove ourselves from the global TLB lock.
            self.cvm_partition().tlb_locked_vps[target_vtl].set_aliased(self_index, false);

            // Check to see whether any other VPs are waiting for this VP to release
            // the TLB lock. Note that other processors may be in the process of
            // inserting themselves into this set because they may have observed that
            // the TLB lock was still held on the current processor, but they will
            // take responsibility for removing themselves after insertion because
            // they will once again observe the TLB lock as not held. Because the set
            // of blocked VPs may be changing, it must be captured locally, since the
            // VP set scan below cannot safely be performed on a VP set that may be
            // changing.
            for blocked_vp in self.cvm_vp_inner().tlb_lock_info[target_vtl]
                .blocked_vps
                .clone()
                .iter_ones()
            {
                self.cvm_vp_inner().tlb_lock_info[target_vtl]
                    .blocked_vps
                    .set_aliased(blocked_vp, false);

                // Mark the target VP as no longer blocked by the current VP.
                // Note that the target VP may have already marked itself as not
                // blocked if is has already noticed that the lock has already
                // been released on the current VP.
                let other_lock = &self
                    .cvm_partition()
                    .vp_inner(blocked_vp as u32)
                    .tlb_lock_info[target_vtl];
                if other_lock.blocking_vps.set_aliased(self_index, false) {
                    let other_old_count =
                        other_lock.blocking_vp_count.fetch_sub(1, Ordering::Relaxed);

                    if other_old_count == 1 {
                        // The current VP was the last one to be removed from the
                        // blocking set of the target VP. If it is asleep, it must
                        // be woken now. Sending an IPI is sufficient to cause it to
                        // reevaluate the blocking state. It is not necessary to
                        // synchronize with its sleep state as a spurious IPI is not
                        // harmful.
                        if other_lock.sleeping.load(Ordering::SeqCst) {
                            self.partition.vps[blocked_vp].wake_vtl2();
                        }
                    }
                }
            }
        }
    }

    /// Unlocks the TLBs of all lower VTLs as required upon VTL exit.
    pub(crate) fn unlock_tlb_lock(&mut self, unlocking_vtl: Vtl) {
        for &target_vtl in &[GuestVtl::Vtl1, GuestVtl::Vtl0][(2 - unlocking_vtl as usize)..] {
            self.unlock_tlb_lock_target(unlocking_vtl, target_vtl);
        }
    }

    /// Returns whether the VP should halt to wait for the TLB lock of the specified VTL.
    pub(crate) fn should_halt_for_tlb_unlock(&mut self, target_vtl: GuestVtl) -> bool {
        // No wait is required unless this VP is blocked on another VP that
        // holds the TLB flush lock.
        let self_lock = &self.cvm_vp_inner().tlb_lock_info[target_vtl];
        if self_lock.blocking_vp_count.load(Ordering::Relaxed) != 0 {
            self_lock.sleeping.store(true, Ordering::Relaxed);
            // Now that this VP has been marked as sleeping, check to see
            // whether it is still blocked. If not, no sleep should be
            // attempted.
            if self_lock.blocking_vp_count.load(Ordering::SeqCst) != 0 {
                return true;
            }

            self_lock.sleeping.store(false, Ordering::Relaxed);
        }

        false
    }
}