@@ -503,10 +503,10 @@ pub(crate) fn codegen_intrinsic_call<'tcx>(
503
503
504
504
if intrinsic == sym:: copy_nonoverlapping {
505
505
// FIXME emit_small_memcpy
506
- fx. bcx. call_memcpy( fx. module . target_config( ) , dst, src, byte_amount) ;
506
+ fx. bcx. call_memcpy( fx. target_config, dst, src, byte_amount) ;
507
507
} else {
508
508
// FIXME emit_small_memmove
509
- fx. bcx. call_memmove( fx. module . target_config( ) , dst, src, byte_amount) ;
509
+ fx. bcx. call_memmove( fx. target_config, dst, src, byte_amount) ;
510
510
}
511
511
} ;
512
512
// NOTE: the volatile variants have src and dst swapped
@@ -522,10 +522,10 @@ pub(crate) fn codegen_intrinsic_call<'tcx>(
522
522
// FIXME make the copy actually volatile when using emit_small_mem{cpy,move}
523
523
if intrinsic == sym:: volatile_copy_nonoverlapping_memory {
524
524
// FIXME emit_small_memcpy
525
- fx. bcx. call_memcpy( fx. module . target_config( ) , dst, src, byte_amount) ;
525
+ fx. bcx. call_memcpy( fx. target_config, dst, src, byte_amount) ;
526
526
} else {
527
527
// FIXME emit_small_memmove
528
- fx. bcx. call_memmove( fx. module . target_config( ) , dst, src, byte_amount) ;
528
+ fx. bcx. call_memmove( fx. target_config, dst, src, byte_amount) ;
529
529
}
530
530
} ;
531
531
size_of_val, <T > ( c ptr) {
@@ -673,7 +673,7 @@ pub(crate) fn codegen_intrinsic_call<'tcx>(
673
673
let dst_ptr = dst. load_scalar( fx) ;
674
674
// FIXME make the memset actually volatile when switching to emit_small_memset
675
675
// FIXME use emit_small_memset
676
- fx. bcx. call_memset( fx. module . target_config( ) , dst_ptr, val, count) ;
676
+ fx. bcx. call_memset( fx. target_config, dst_ptr, val, count) ;
677
677
} ;
678
678
ctlz | ctlz_nonzero, <T > ( v arg) {
679
679
// FIXME trap on `ctlz_nonzero` with zero arg.
0 commit comments