1 | //! A fixed capacity Multiple-Producer Multiple-Consumer (MPMC) lock-free queue |
2 | //! |
3 | //! NOTE: This module requires atomic CAS operations. On targets where they're not natively available, |
4 | //! they are emulated by the [`portable-atomic`](https://crates.io/crates/portable-atomic) crate. |
5 | //! |
6 | //! # Example |
7 | //! |
8 | //! This queue can be constructed in "const context". Placing it in a `static` variable lets *all* |
9 | //! contexts (interrupts / threads / `main`) safely enqueue and dequeue items from it. |
10 | //! |
11 | //! ``` ignore |
12 | //! #![no_main] |
13 | //! #![no_std] |
14 | //! |
15 | //! use panic_semihosting as _; |
16 | //! |
17 | //! use cortex_m::{asm, peripheral::syst::SystClkSource}; |
18 | //! use cortex_m_rt::{entry, exception}; |
19 | //! use cortex_m_semihosting::hprintln; |
20 | //! use heapless::mpmc::Q2; |
21 | //! |
22 | //! static Q: Q2<u8> = Q2::new(); |
23 | //! |
24 | //! #[entry] |
25 | //! fn main() -> ! { |
26 | //! if let Some(p) = cortex_m::Peripherals::take() { |
27 | //! let mut syst = p.SYST; |
28 | //! |
29 | //! // configures the system timer to trigger a SysTick exception every second |
30 | //! syst.set_clock_source(SystClkSource::Core); |
31 | //! syst.set_reload(12_000_000); |
32 | //! syst.enable_counter(); |
33 | //! syst.enable_interrupt(); |
34 | //! } |
35 | //! |
36 | //! loop { |
37 | //! if let Some(x) = Q.dequeue() { |
38 | //! hprintln!("{}" , x).ok(); |
39 | //! } else { |
40 | //! asm::wfi(); |
41 | //! } |
42 | //! } |
43 | //! } |
44 | //! |
45 | //! #[exception] |
46 | //! fn SysTick() { |
47 | //! static mut COUNT: u8 = 0; |
48 | //! |
49 | //! Q.enqueue(*COUNT).ok(); |
50 | //! *COUNT += 1; |
51 | //! } |
52 | //! ``` |
53 | //! |
54 | //! # Benchmark |
55 | //! |
56 | //! Measured on a ARM Cortex-M3 core running at 8 MHz and with zero Flash wait cycles |
57 | //! |
58 | //! N| `Q8::<u8>::enqueue().ok()` (`z`) | `Q8::<u8>::dequeue()` (`z`) | |
59 | //! -|----------------------------------|-----------------------------| |
60 | //! 0|34 |35 | |
61 | //! 1|52 |53 | |
62 | //! 2|69 |71 | |
63 | //! |
64 | //! - `N` denotes the number of *interruptions*. On Cortex-M, an interruption consists of an |
65 | //! interrupt handler preempting the would-be atomic section of the `enqueue` / `dequeue` |
66 | //! operation. Note that it does *not* matter if the higher priority handler uses the queue or |
67 | //! not. |
68 | //! - All execution times are in clock cycles. 1 clock cycle = 125 ns. |
69 | //! - Execution time is *dependent* of `mem::size_of::<T>()`. Both operations include one |
70 | //! `memcpy(T)` in their successful path. |
71 | //! - The optimization level is indicated in parentheses. |
72 | //! - The numbers reported correspond to the successful path (i.e. `Some` is returned by `dequeue` |
73 | //! and `Ok` is returned by `enqueue`). |
74 | //! |
75 | //! # Portability |
76 | //! |
77 | //! This module requires CAS atomic instructions which are not available on all architectures |
78 | //! (e.g. ARMv6-M (`thumbv6m-none-eabi`) and MSP430 (`msp430-none-elf`)). These atomics can be |
79 | //! emulated however with [`portable-atomic`](https://crates.io/crates/portable-atomic), which is |
80 | //! enabled with the `cas` feature and is enabled by default for `thumbv6m-none-eabi` and `riscv32` |
81 | //! targets. |
82 | //! |
83 | //! # References |
84 | //! |
85 | //! This is an implementation of Dmitry Vyukov's ["Bounded MPMC queue"][0] minus the cache padding. |
86 | //! |
87 | //! [0]: http://www.1024cores.net/home/lock-free-algorithms/queues/bounded-mpmc-queue |
88 | |
89 | use core::{cell::UnsafeCell, mem::MaybeUninit}; |
90 | |
91 | #[cfg (not(feature = "portable-atomic" ))] |
92 | use core::sync::atomic; |
93 | #[cfg (feature = "portable-atomic" )] |
94 | use portable_atomic as atomic; |
95 | |
96 | use atomic::Ordering; |
97 | |
98 | #[cfg (feature = "mpmc_large" )] |
99 | type AtomicTargetSize = atomic::AtomicUsize; |
100 | #[cfg (not(feature = "mpmc_large" ))] |
101 | type AtomicTargetSize = atomic::AtomicU8; |
102 | |
103 | #[cfg (feature = "mpmc_large" )] |
104 | type IntSize = usize; |
105 | #[cfg (not(feature = "mpmc_large" ))] |
106 | type IntSize = u8; |
107 | |
108 | /// MPMC queue with a capability for 2 elements. |
109 | pub type Q2<T> = MpMcQueue<T, 2>; |
110 | |
111 | /// MPMC queue with a capability for 4 elements. |
112 | pub type Q4<T> = MpMcQueue<T, 4>; |
113 | |
114 | /// MPMC queue with a capability for 8 elements. |
115 | pub type Q8<T> = MpMcQueue<T, 8>; |
116 | |
117 | /// MPMC queue with a capability for 16 elements. |
118 | pub type Q16<T> = MpMcQueue<T, 16>; |
119 | |
120 | /// MPMC queue with a capability for 32 elements. |
121 | pub type Q32<T> = MpMcQueue<T, 32>; |
122 | |
123 | /// MPMC queue with a capability for 64 elements. |
124 | pub type Q64<T> = MpMcQueue<T, 64>; |
125 | |
126 | /// MPMC queue with a capacity for N elements |
127 | /// N must be a power of 2 |
128 | /// The max value of N is u8::MAX - 1 if `mpmc_large` feature is not enabled. |
129 | pub struct MpMcQueue<T, const N: usize> { |
130 | buffer: UnsafeCell<[Cell<T>; N]>, |
131 | dequeue_pos: AtomicTargetSize, |
132 | enqueue_pos: AtomicTargetSize, |
133 | } |
134 | |
135 | impl<T, const N: usize> MpMcQueue<T, N> { |
136 | const MASK: IntSize = (N - 1) as IntSize; |
137 | const EMPTY_CELL: Cell<T> = Cell::new(0); |
138 | |
139 | const ASSERT: [(); 1] = [()]; |
140 | |
141 | /// Creates an empty queue |
142 | pub const fn new() -> Self { |
143 | // Const assert |
144 | crate::sealed::greater_than_1::<N>(); |
145 | crate::sealed::power_of_two::<N>(); |
146 | |
147 | // Const assert on size. |
148 | Self::ASSERT[!(N < (IntSize::MAX as usize)) as usize]; |
149 | |
150 | let mut cell_count = 0; |
151 | |
152 | let mut result_cells: [Cell<T>; N] = [Self::EMPTY_CELL; N]; |
153 | while cell_count != N { |
154 | result_cells[cell_count] = Cell::new(cell_count); |
155 | cell_count += 1; |
156 | } |
157 | |
158 | Self { |
159 | buffer: UnsafeCell::new(result_cells), |
160 | dequeue_pos: AtomicTargetSize::new(0), |
161 | enqueue_pos: AtomicTargetSize::new(0), |
162 | } |
163 | } |
164 | |
165 | /// Returns the item in the front of the queue, or `None` if the queue is empty |
166 | pub fn dequeue(&self) -> Option<T> { |
167 | unsafe { dequeue(self.buffer.get() as *mut _, &self.dequeue_pos, Self::MASK) } |
168 | } |
169 | |
170 | /// Adds an `item` to the end of the queue |
171 | /// |
172 | /// Returns back the `item` if the queue is full |
173 | pub fn enqueue(&self, item: T) -> Result<(), T> { |
174 | unsafe { |
175 | enqueue( |
176 | self.buffer.get() as *mut _, |
177 | &self.enqueue_pos, |
178 | Self::MASK, |
179 | item, |
180 | ) |
181 | } |
182 | } |
183 | } |
184 | |
185 | impl<T, const N: usize> Default for MpMcQueue<T, N> { |
186 | fn default() -> Self { |
187 | Self::new() |
188 | } |
189 | } |
190 | |
191 | unsafe impl<T, const N: usize> Sync for MpMcQueue<T, N> where T: Send {} |
192 | |
193 | struct Cell<T> { |
194 | data: MaybeUninit<T>, |
195 | sequence: AtomicTargetSize, |
196 | } |
197 | |
198 | impl<T> Cell<T> { |
199 | const fn new(seq: usize) -> Self { |
200 | Self { |
201 | data: MaybeUninit::uninit(), |
202 | sequence: AtomicTargetSize::new(seq as IntSize), |
203 | } |
204 | } |
205 | } |
206 | |
207 | unsafe fn dequeue<T>( |
208 | buffer: *mut Cell<T>, |
209 | dequeue_pos: &AtomicTargetSize, |
210 | mask: IntSize, |
211 | ) -> Option<T> { |
212 | let mut pos = dequeue_pos.load(Ordering::Relaxed); |
213 | |
214 | let mut cell; |
215 | loop { |
216 | cell = buffer.add(usize::from(pos & mask)); |
217 | let seq = (*cell).sequence.load(Ordering::Acquire); |
218 | let dif = (seq as i8).wrapping_sub((pos.wrapping_add(1)) as i8); |
219 | |
220 | if dif == 0 { |
221 | if dequeue_pos |
222 | .compare_exchange_weak( |
223 | pos, |
224 | pos.wrapping_add(1), |
225 | Ordering::Relaxed, |
226 | Ordering::Relaxed, |
227 | ) |
228 | .is_ok() |
229 | { |
230 | break; |
231 | } |
232 | } else if dif < 0 { |
233 | return None; |
234 | } else { |
235 | pos = dequeue_pos.load(Ordering::Relaxed); |
236 | } |
237 | } |
238 | |
239 | let data = (*cell).data.as_ptr().read(); |
240 | (*cell) |
241 | .sequence |
242 | .store(pos.wrapping_add(mask).wrapping_add(1), Ordering::Release); |
243 | Some(data) |
244 | } |
245 | |
246 | unsafe fn enqueue<T>( |
247 | buffer: *mut Cell<T>, |
248 | enqueue_pos: &AtomicTargetSize, |
249 | mask: IntSize, |
250 | item: T, |
251 | ) -> Result<(), T> { |
252 | let mut pos = enqueue_pos.load(Ordering::Relaxed); |
253 | |
254 | let mut cell; |
255 | loop { |
256 | cell = buffer.add(usize::from(pos & mask)); |
257 | let seq = (*cell).sequence.load(Ordering::Acquire); |
258 | let dif = (seq as i8).wrapping_sub(pos as i8); |
259 | |
260 | if dif == 0 { |
261 | if enqueue_pos |
262 | .compare_exchange_weak( |
263 | pos, |
264 | pos.wrapping_add(1), |
265 | Ordering::Relaxed, |
266 | Ordering::Relaxed, |
267 | ) |
268 | .is_ok() |
269 | { |
270 | break; |
271 | } |
272 | } else if dif < 0 { |
273 | return Err(item); |
274 | } else { |
275 | pos = enqueue_pos.load(Ordering::Relaxed); |
276 | } |
277 | } |
278 | |
279 | (*cell).data.as_mut_ptr().write(item); |
280 | (*cell) |
281 | .sequence |
282 | .store(pos.wrapping_add(1), Ordering::Release); |
283 | Ok(()) |
284 | } |
285 | |
286 | #[cfg (test)] |
287 | mod tests { |
288 | use super::Q2; |
289 | |
290 | #[test ] |
291 | fn sanity() { |
292 | let q = Q2::new(); |
293 | q.enqueue(0).unwrap(); |
294 | q.enqueue(1).unwrap(); |
295 | assert!(q.enqueue(2).is_err()); |
296 | |
297 | assert_eq!(q.dequeue(), Some(0)); |
298 | assert_eq!(q.dequeue(), Some(1)); |
299 | assert_eq!(q.dequeue(), None); |
300 | } |
301 | |
302 | #[test ] |
303 | fn drain_at_pos255() { |
304 | let q = Q2::new(); |
305 | for _ in 0..255 { |
306 | assert!(q.enqueue(0).is_ok()); |
307 | assert_eq!(q.dequeue(), Some(0)); |
308 | } |
309 | // this should not block forever |
310 | assert_eq!(q.dequeue(), None); |
311 | } |
312 | |
313 | #[test ] |
314 | fn full_at_wrapped_pos0() { |
315 | let q = Q2::new(); |
316 | for _ in 0..254 { |
317 | assert!(q.enqueue(0).is_ok()); |
318 | assert_eq!(q.dequeue(), Some(0)); |
319 | } |
320 | assert!(q.enqueue(0).is_ok()); |
321 | assert!(q.enqueue(0).is_ok()); |
322 | // this should not block forever |
323 | assert!(q.enqueue(0).is_err()); |
324 | } |
325 | } |
326 | |