|
1 | 1 | use super::abi::AbiBuilderMethods;
|
2 | 2 | use super::asm::AsmBuilderMethods;
|
| 3 | +use super::consts::ConstMethods; |
3 | 4 | use super::coverageinfo::CoverageInfoBuilderMethods;
|
4 | 5 | use super::debuginfo::DebugInfoBuilderMethods;
|
5 | 6 | use super::intrinsic::IntrinsicCallMethods;
|
6 | 7 | use super::misc::MiscMethods;
|
7 |
| -use super::type_::{ArgAbiMethods, BaseTypeMethods}; |
| 8 | +use super::type_::{ArgAbiMethods, BaseTypeMethods, LayoutTypeMethods}; |
8 | 9 | use super::{HasCodegen, StaticBuilderMethods};
|
9 | 10 |
|
10 | 11 | use crate::common::{
|
11 | 12 | AtomicOrdering, AtomicRmwBinOp, IntPredicate, RealPredicate, SynchronizationScope, TypeKind,
|
12 | 13 | };
|
13 |
| -use crate::mir::operand::OperandRef; |
| 14 | +use crate::mir::operand::{OperandRef, OperandValue}; |
14 | 15 | use crate::mir::place::PlaceRef;
|
15 | 16 | use crate::MemFlags;
|
16 | 17 |
|
17 | 18 | use rustc_middle::middle::codegen_fn_attrs::CodegenFnAttrs;
|
18 | 19 | use rustc_middle::ty::layout::{HasParamEnv, TyAndLayout};
|
19 | 20 | use rustc_middle::ty::Ty;
|
| 21 | +use rustc_session::config::OptLevel; |
20 | 22 | use rustc_span::Span;
|
21 | 23 | use rustc_target::abi::call::FnAbi;
|
22 | 24 | use rustc_target::abi::{Abi, Align, Scalar, Size, WrappingRange};
|
@@ -267,6 +269,54 @@ pub trait BuilderMethods<'a, 'tcx>:
|
267 | 269 | flags: MemFlags,
|
268 | 270 | );
|
269 | 271 |
|
| 272 | + /// *Typed* copy for non-overlapping places. |
| 273 | + /// |
| 274 | + /// Has a default implementation in terms of `memcpy`, but specific backends |
| 275 | + /// can override to do something smarter if possible. |
| 276 | + /// |
| 277 | + /// (For example, typed load-stores with alias metadata.) |
| 278 | + fn typed_place_copy( |
| 279 | + &mut self, |
| 280 | + dst: PlaceRef<'tcx, Self::Value>, |
| 281 | + src: PlaceRef<'tcx, Self::Value>, |
| 282 | + ) { |
| 283 | + debug_assert!(src.llextra.is_none()); |
| 284 | + debug_assert!(dst.llextra.is_none()); |
| 285 | + debug_assert_eq!(dst.layout.size, src.layout.size); |
| 286 | + if self.sess().opts.optimize == OptLevel::No && self.is_backend_immediate(dst.layout) { |
| 287 | + // If we're not optimizing, the aliasing information from `memcpy` |
| 288 | + // isn't useful, so just load-store the value for smaller code. |
| 289 | + let temp = self.load_operand(src); |
| 290 | + temp.val.store(self, dst); |
| 291 | + } else if !dst.layout.is_zst() { |
| 292 | + let bytes = self.const_usize(dst.layout.size.bytes()); |
| 293 | + self.memcpy(dst.llval, dst.align, src.llval, src.align, bytes, MemFlags::empty()); |
| 294 | + } |
| 295 | + } |
| 296 | + |
| 297 | + /// *Typed* swap for non-overlapping places. |
| 298 | + /// |
| 299 | + /// Avoids `alloca`s for Immediates and ScalarPairs. |
| 300 | + /// |
| 301 | + /// FIXME: Maybe do something smarter for Ref types too? |
| 302 | + /// For now, the `typed_swap` intrinsic just doesn't call this for those |
| 303 | + /// cases (in non-debug), preferring the fallback body instead. |
| 304 | + fn typed_place_swap( |
| 305 | + &mut self, |
| 306 | + left: PlaceRef<'tcx, Self::Value>, |
| 307 | + right: PlaceRef<'tcx, Self::Value>, |
| 308 | + ) { |
| 309 | + let mut temp = self.load_operand(left); |
| 310 | + if let OperandValue::Ref(..) = temp.val { |
| 311 | + // The SSA value isn't stand-alone, so we need to copy it elsewhere |
| 312 | + let alloca = PlaceRef::alloca(self, left.layout); |
| 313 | + self.typed_place_copy(alloca, left); |
| 314 | + temp = self.load_operand(alloca); |
| 315 | + } |
| 316 | + self.typed_place_copy(left, right); |
| 317 | + temp.val.store(self, right); |
| 318 | + } |
| 319 | + |
270 | 320 | fn select(
|
271 | 321 | &mut self,
|
272 | 322 | cond: Self::Value,
|
|
0 commit comments