Skip to content

Commit fecb7b4

Browse files
committed
Auto merge of #124193 - RalfJung:miri, r=RalfJung
Miri subtree update r? `@ghost`
2 parents b9be3c4 + ae37b6e commit fecb7b4

File tree

114 files changed

+1400
-685
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

114 files changed

+1400
-685
lines changed

src/tools/miri/README.md

+10
Original file line numberDiff line numberDiff line change
@@ -295,6 +295,16 @@ up the sysroot. If you are using `miri` (the Miri driver) directly, see the
295295
Miri adds its own set of `-Z` flags, which are usually set via the `MIRIFLAGS`
296296
environment variable. We first document the most relevant and most commonly used flags:
297297

298+
* `-Zmiri-address-reuse-rate=<rate>` changes the probability that a freed *non-stack* allocation
299+
will be added to the pool for address reuse, and the probability that a new *non-stack* allocation
300+
will be taken from the pool. Stack allocations never get added to or taken from the pool. The
301+
default is `0.5`.
302+
* `-Zmiri-address-reuse-cross-thread-rate=<rate>` changes the probability that an allocation which
303+
attempts to reuse a previously freed block of memory will also consider blocks freed by *other
304+
threads*. The default is `0.1`, which means by default, in 90% of the cases where an address reuse
305+
attempt is made, only addresses from the same thread will be considered. Reusing an address from
306+
another thread induces synchronization between those threads, which can mask data races and weak
307+
memory bugs.
298308
* `-Zmiri-compare-exchange-weak-failure-rate=<rate>` changes the failure rate of
299309
`compare_exchange_weak` operations. The default is `0.8` (so 4 out of 5 weak ops will fail).
300310
You can change it to any value between `0.0` and `1.0`, where `1.0` means it

src/tools/miri/rust-version

+1-1
Original file line numberDiff line numberDiff line change
@@ -1 +1 @@
1-
23d47dba319331d4418827cfbb8c1af283497d3c
1+
c8d19a92aa9022eb690899cf6d54fd23cb6877e5

src/tools/miri/src/alloc_addresses/mod.rs

+37-20
Original file line numberDiff line numberDiff line change
@@ -13,8 +13,9 @@ use rustc_data_structures::fx::{FxHashMap, FxHashSet};
1313
use rustc_span::Span;
1414
use rustc_target::abi::{Align, HasDataLayout, Size};
1515

16-
use crate::*;
17-
use reuse_pool::ReusePool;
16+
use crate::{concurrency::VClock, *};
17+
18+
use self::reuse_pool::ReusePool;
1819

1920
#[derive(Copy, Clone, Debug, PartialEq, Eq)]
2021
pub enum ProvenanceMode {
@@ -77,7 +78,7 @@ impl GlobalStateInner {
7778
GlobalStateInner {
7879
int_to_ptr_map: Vec::default(),
7980
base_addr: FxHashMap::default(),
80-
reuse: ReusePool::new(),
81+
reuse: ReusePool::new(config),
8182
exposed: FxHashSet::default(),
8283
next_base_addr: stack_addr,
8384
provenance_mode: config.provenance_mode,
@@ -144,7 +145,7 @@ trait EvalContextExtPriv<'mir, 'tcx: 'mir>: crate::MiriInterpCxExt<'mir, 'tcx> {
144145
fn addr_from_alloc_id(
145146
&self,
146147
alloc_id: AllocId,
147-
_kind: MemoryKind,
148+
memory_kind: MemoryKind,
148149
) -> InterpResult<'tcx, u64> {
149150
let ecx = self.eval_context_ref();
150151
let mut global_state = ecx.machine.alloc_addresses.borrow_mut();
@@ -163,9 +164,18 @@ trait EvalContextExtPriv<'mir, 'tcx: 'mir>: crate::MiriInterpCxExt<'mir, 'tcx> {
163164
assert!(!matches!(kind, AllocKind::Dead));
164165

165166
// This allocation does not have a base address yet, pick or reuse one.
166-
let base_addr = if let Some(reuse_addr) =
167-
global_state.reuse.take_addr(&mut *rng, size, align)
168-
{
167+
let base_addr = if let Some((reuse_addr, clock)) = global_state.reuse.take_addr(
168+
&mut *rng,
169+
size,
170+
align,
171+
memory_kind,
172+
ecx.get_active_thread(),
173+
) {
174+
if let Some(clock) = clock
175+
&& let Some(data_race) = &ecx.machine.data_race
176+
{
177+
data_race.acquire_clock(&clock, ecx.get_active_thread());
178+
}
169179
reuse_addr
170180
} else {
171181
// We have to pick a fresh address.
@@ -333,14 +343,11 @@ pub trait EvalContextExt<'mir, 'tcx: 'mir>: crate::MiriInterpCxExt<'mir, 'tcx> {
333343
}
334344
}
335345

336-
impl GlobalStateInner {
337-
pub fn free_alloc_id(
338-
&mut self,
339-
rng: &mut impl Rng,
340-
dead_id: AllocId,
341-
size: Size,
342-
align: Align,
343-
) {
346+
impl<'mir, 'tcx> MiriMachine<'mir, 'tcx> {
347+
pub fn free_alloc_id(&mut self, dead_id: AllocId, size: Size, align: Align, kind: MemoryKind) {
348+
let global_state = self.alloc_addresses.get_mut();
349+
let rng = self.rng.get_mut();
350+
344351
// We can *not* remove this from `base_addr`, since the interpreter design requires that we
345352
// be able to retrieve an AllocId + offset for any memory access *before* we check if the
346353
// access is valid. Specifically, `ptr_get_alloc` is called on each attempt at a memory
@@ -353,15 +360,25 @@ impl GlobalStateInner {
353360
// returns a dead allocation.
354361
// To avoid a linear scan we first look up the address in `base_addr`, and then find it in
355362
// `int_to_ptr_map`.
356-
let addr = *self.base_addr.get(&dead_id).unwrap();
357-
let pos = self.int_to_ptr_map.binary_search_by_key(&addr, |(addr, _)| *addr).unwrap();
358-
let removed = self.int_to_ptr_map.remove(pos);
363+
let addr = *global_state.base_addr.get(&dead_id).unwrap();
364+
let pos =
365+
global_state.int_to_ptr_map.binary_search_by_key(&addr, |(addr, _)| *addr).unwrap();
366+
let removed = global_state.int_to_ptr_map.remove(pos);
359367
assert_eq!(removed, (addr, dead_id)); // double-check that we removed the right thing
360368
// We can also remove it from `exposed`, since this allocation can anyway not be returned by
361369
// `alloc_id_from_addr` any more.
362-
self.exposed.remove(&dead_id);
370+
global_state.exposed.remove(&dead_id);
363371
// Also remember this address for future reuse.
364-
self.reuse.add_addr(rng, addr, size, align)
372+
let thread = self.threads.get_active_thread_id();
373+
global_state.reuse.add_addr(rng, addr, size, align, kind, thread, || {
374+
if let Some(data_race) = &self.data_race {
375+
data_race
376+
.release_clock(thread, self.threads.active_thread_ref().current_span())
377+
.clone()
378+
} else {
379+
VClock::default()
380+
}
381+
})
365382
}
366383
}
367384

src/tools/miri/src/alloc_addresses/reuse_pool.rs

+64-22
Original file line numberDiff line numberDiff line change
@@ -4,73 +4,113 @@ use rand::Rng;
44

55
use rustc_target::abi::{Align, Size};
66

7-
const MAX_POOL_SIZE: usize = 64;
7+
use crate::{concurrency::VClock, MemoryKind, MiriConfig, ThreadId};
88

9-
// Just use fair coins, until we have evidence that other numbers are better.
10-
const ADDR_REMEMBER_CHANCE: f64 = 0.5;
11-
const ADDR_TAKE_CHANCE: f64 = 0.5;
9+
const MAX_POOL_SIZE: usize = 64;
1210

1311
/// The pool strikes a balance between exploring more possible executions and making it more likely
1412
/// to find bugs. The hypothesis is that bugs are more likely to occur when reuse happens for
1513
/// allocations with the same layout, since that can trigger e.g. ABA issues in a concurrent data
1614
/// structure. Therefore we only reuse allocations when size and alignment match exactly.
1715
#[derive(Debug)]
1816
pub struct ReusePool {
17+
address_reuse_rate: f64,
18+
address_reuse_cross_thread_rate: f64,
1919
/// The i-th element in `pool` stores allocations of alignment `2^i`. We store these reusable
20-
/// allocations as address-size pairs, the list must be sorted by the size.
20+
/// allocations as address-size pairs, the list must be sorted by the size and then the thread ID.
2121
///
2222
/// Each of these maps has at most MAX_POOL_SIZE elements, and since alignment is limited to
2323
/// less than 64 different possible value, that bounds the overall size of the pool.
24-
pool: Vec<Vec<(u64, Size)>>,
24+
///
25+
/// We also store the ID and the data-race clock of the thread that donated this pool element,
26+
/// to ensure synchronization with the thread that picks up this address.
27+
pool: Vec<Vec<(u64, Size, ThreadId, VClock)>>,
2528
}
2629

2730
impl ReusePool {
28-
pub fn new() -> Self {
29-
ReusePool { pool: vec![] }
31+
pub fn new(config: &MiriConfig) -> Self {
32+
ReusePool {
33+
address_reuse_rate: config.address_reuse_rate,
34+
address_reuse_cross_thread_rate: config.address_reuse_cross_thread_rate,
35+
pool: vec![],
36+
}
3037
}
3138

32-
fn subpool(&mut self, align: Align) -> &mut Vec<(u64, Size)> {
39+
fn subpool(&mut self, align: Align) -> &mut Vec<(u64, Size, ThreadId, VClock)> {
3340
let pool_idx: usize = align.bytes().trailing_zeros().try_into().unwrap();
3441
if self.pool.len() <= pool_idx {
3542
self.pool.resize(pool_idx + 1, Vec::new());
3643
}
3744
&mut self.pool[pool_idx]
3845
}
3946

40-
pub fn add_addr(&mut self, rng: &mut impl Rng, addr: u64, size: Size, align: Align) {
47+
pub fn add_addr(
48+
&mut self,
49+
rng: &mut impl Rng,
50+
addr: u64,
51+
size: Size,
52+
align: Align,
53+
kind: MemoryKind,
54+
thread: ThreadId,
55+
clock: impl FnOnce() -> VClock,
56+
) {
4157
// Let's see if we even want to remember this address.
42-
if !rng.gen_bool(ADDR_REMEMBER_CHANCE) {
58+
// We don't remember stack addresses: there's a lot of them (so the perf impact is big),
59+
// and we only want to reuse stack slots within the same thread or else we'll add a lot of
60+
// undesired synchronization.
61+
if kind == MemoryKind::Stack || !rng.gen_bool(self.address_reuse_rate) {
4362
return;
4463
}
64+
let clock = clock();
4565
// Determine the pool to add this to, and where in the pool to put it.
4666
let subpool = self.subpool(align);
47-
let pos = subpool.partition_point(|(_addr, other_size)| *other_size < size);
67+
let pos = subpool.partition_point(|(_addr, other_size, other_thread, _)| {
68+
(*other_size, *other_thread) < (size, thread)
69+
});
4870
// Make sure the pool does not grow too big.
4971
if subpool.len() >= MAX_POOL_SIZE {
5072
// Pool full. Replace existing element, or last one if this would be even bigger.
5173
let clamped_pos = pos.min(subpool.len() - 1);
52-
subpool[clamped_pos] = (addr, size);
74+
subpool[clamped_pos] = (addr, size, thread, clock);
5375
return;
5476
}
5577
// Add address to pool, at the right position.
56-
subpool.insert(pos, (addr, size));
78+
subpool.insert(pos, (addr, size, thread, clock));
5779
}
5880

59-
pub fn take_addr(&mut self, rng: &mut impl Rng, size: Size, align: Align) -> Option<u64> {
60-
// Determine whether we'll even attempt a reuse.
61-
if !rng.gen_bool(ADDR_TAKE_CHANCE) {
81+
/// Returns the address to use and optionally a clock we have to synchronize with.
82+
pub fn take_addr(
83+
&mut self,
84+
rng: &mut impl Rng,
85+
size: Size,
86+
align: Align,
87+
kind: MemoryKind,
88+
thread: ThreadId,
89+
) -> Option<(u64, Option<VClock>)> {
90+
// Determine whether we'll even attempt a reuse. As above, we don't do reuse for stack addresses.
91+
if kind == MemoryKind::Stack || !rng.gen_bool(self.address_reuse_rate) {
6292
return None;
6393
}
94+
let cross_thread_reuse = rng.gen_bool(self.address_reuse_cross_thread_rate);
6495
// Determine the pool to take this from.
6596
let subpool = self.subpool(align);
6697
// Let's see if we can find something of the right size. We want to find the full range of
67-
// such items, beginning with the first, so we can't use `binary_search_by_key`.
68-
let begin = subpool.partition_point(|(_addr, other_size)| *other_size < size);
98+
// such items, beginning with the first, so we can't use `binary_search_by_key`. If we do
99+
// *not* want to consider other thread's allocations, we effectively use the lexicographic
100+
// order on `(size, thread)`.
101+
let begin = subpool.partition_point(|(_addr, other_size, other_thread, _)| {
102+
*other_size < size
103+
|| (*other_size == size && !cross_thread_reuse && *other_thread < thread)
104+
});
69105
let mut end = begin;
70-
while let Some((_addr, other_size)) = subpool.get(end) {
106+
while let Some((_addr, other_size, other_thread, _)) = subpool.get(end) {
71107
if *other_size != size {
72108
break;
73109
}
110+
if !cross_thread_reuse && *other_thread != thread {
111+
// We entered the allocations of another thread.
112+
break;
113+
}
74114
end += 1;
75115
}
76116
if end == begin {
@@ -80,8 +120,10 @@ impl ReusePool {
80120
// Pick a random element with the desired size.
81121
let idx = rng.gen_range(begin..end);
82122
// Remove it from the pool and return.
83-
let (chosen_addr, chosen_size) = subpool.remove(idx);
123+
let (chosen_addr, chosen_size, chosen_thread, clock) = subpool.remove(idx);
84124
debug_assert!(chosen_size >= size && chosen_addr % align.bytes() == 0);
85-
Some(chosen_addr)
125+
debug_assert!(cross_thread_reuse || chosen_thread == thread);
126+
// No synchronization needed if we reused from the current thread.
127+
Some((chosen_addr, if chosen_thread == thread { None } else { Some(clock) }))
86128
}
87129
}

0 commit comments

Comments
 (0)