9
9
// except according to those terms.
10
10
11
11
use alloc:: heap:: { allocate, deallocate, EMPTY } ;
12
+ use core_alloc:: Allocator ;
13
+ use core_alloc:: heap:: Allocator as HeapAllocator ;
12
14
13
15
use cmp;
14
16
use hash:: { Hash , Hasher } ;
@@ -61,14 +63,15 @@ const EMPTY_BUCKET: u64 = 0;
61
63
/// invariants at the type level and employs some performance trickery,
62
64
/// but in general is just a tricked out `Vec<Option<u64, K, V>>`.
63
65
#[ unsafe_no_drop_flag]
64
- pub struct RawTable < K , V > {
66
+ pub struct RawTable < K , V , A = HeapAllocator > where A : Allocator {
65
67
capacity : usize ,
66
68
size : usize ,
67
69
hashes : Unique < u64 > ,
68
70
69
71
// Because K/V do not appear directly in any of the types in the struct,
70
72
// inform rustc that in fact instances of K and V are reachable from here.
71
73
marker : marker:: PhantomData < ( K , V ) > ,
74
+ alloc : A ,
72
75
}
73
76
74
77
unsafe impl < K : Send , V : Send > Send for RawTable < K , V > { }
@@ -219,7 +222,7 @@ impl<K, V, M> Bucket<K, V, M> {
219
222
}
220
223
}
221
224
222
- impl < K , V , M : Deref < Target =RawTable < K , V > > > Bucket < K , V , M > {
225
+ impl < K , V , A , M : Deref < Target =RawTable < K , V , A > > > Bucket < K , V , M > where A : Allocator {
223
226
pub fn new ( table : M , hash : SafeHash ) -> Bucket < K , V , M > {
224
227
Bucket :: at_index ( table, hash. inspect ( ) as usize )
225
228
}
@@ -291,7 +294,7 @@ impl<K, V, M: Deref<Target=RawTable<K, V>>> Bucket<K, V, M> {
291
294
}
292
295
}
293
296
294
- impl < K , V , M : Deref < Target =RawTable < K , V > > > EmptyBucket < K , V , M > {
297
+ impl < K , V , A , M : Deref < Target =RawTable < K , V , A > > > EmptyBucket < K , V , M > where A : Allocator {
295
298
#[ inline]
296
299
pub fn next ( self ) -> Bucket < K , V , M > {
297
300
let mut bucket = self . into_bucket ( ) ;
@@ -349,7 +352,7 @@ impl<K, V, M: Deref<Target=RawTable<K, V>> + DerefMut> EmptyBucket<K, V, M> {
349
352
}
350
353
}
351
354
352
- impl < K , V , M : Deref < Target =RawTable < K , V > > > FullBucket < K , V , M > {
355
+ impl < K , V , A , M : Deref < Target =RawTable < K , V , A > > > FullBucket < K , V , M > where A : Allocator {
353
356
#[ inline]
354
357
pub fn next ( self ) -> Bucket < K , V , M > {
355
358
let mut bucket = self . into_bucket ( ) ;
@@ -567,15 +570,24 @@ fn test_offset_calculation() {
567
570
}
568
571
569
572
impl < K , V > RawTable < K , V > {
573
+ /// Creates a new raw table from a given capacity. All buckets are
574
+ /// initially empty.
575
+ pub fn new ( capacity : usize ) -> RawTable < K , V > {
576
+ RawTable :: new_in ( capacity, HeapAllocator )
577
+ }
578
+ }
579
+
580
+ impl < K , V , A > RawTable < K , V , A > where A : Allocator {
570
581
/// Does not initialize the buckets. The caller should ensure they,
571
582
/// at the very least, set every hash to EMPTY_BUCKET.
572
- unsafe fn new_uninitialized ( capacity : usize ) -> RawTable < K , V > {
583
+ unsafe fn new_uninitialized_in ( capacity : usize , a : A ) -> RawTable < K , V , A > {
573
584
if capacity == 0 {
574
585
return RawTable {
575
586
size : 0 ,
576
587
capacity : 0 ,
577
588
hashes : Unique :: new ( EMPTY as * mut u64 ) ,
578
589
marker : marker:: PhantomData ,
590
+ alloc : a,
579
591
} ;
580
592
}
581
593
@@ -618,6 +630,7 @@ impl<K, V> RawTable<K, V> {
618
630
size : 0 ,
619
631
hashes : Unique :: new ( hashes) ,
620
632
marker : marker:: PhantomData ,
633
+ alloc : a,
621
634
}
622
635
}
623
636
@@ -641,11 +654,10 @@ impl<K, V> RawTable<K, V> {
641
654
}
642
655
}
643
656
644
- /// Creates a new raw table from a given capacity. All buckets are
645
- /// initially empty.
646
- pub fn new ( capacity : usize ) -> RawTable < K , V > {
657
+ /// DOC TODO
658
+ pub fn new_in ( capacity : usize , a : A ) -> RawTable < K , V , A > {
647
659
unsafe {
648
- let ret = RawTable :: new_uninitialized ( capacity) ;
660
+ let ret = RawTable :: new_uninitialized_in ( capacity, a ) ;
649
661
ptr:: write_bytes ( * ret. hashes , 0 , capacity) ;
650
662
ret
651
663
}
@@ -686,7 +698,7 @@ impl<K, V> RawTable<K, V> {
686
698
}
687
699
}
688
700
689
- pub fn into_iter ( self ) -> IntoIter < K , V > {
701
+ pub fn into_iter ( self ) -> IntoIter < K , V , A > {
690
702
let RawBuckets { raw, hashes_end, .. } = self . raw_buckets ( ) ;
691
703
// Replace the marker regardless of lifetime bounds on parameters.
692
704
IntoIter {
@@ -699,7 +711,7 @@ impl<K, V> RawTable<K, V> {
699
711
}
700
712
}
701
713
702
- pub fn drain ( & mut self ) -> Drain < K , V > {
714
+ pub fn drain ( & mut self ) -> Drain < K , V , A > {
703
715
let RawBuckets { raw, hashes_end, .. } = self . raw_buckets ( ) ;
704
716
// Replace the marker regardless of lifetime bounds on parameters.
705
717
Drain {
@@ -842,22 +854,22 @@ unsafe impl<'a, K: Sync, V: Sync> Sync for IterMut<'a, K, V> {}
842
854
unsafe impl < ' a , K : Send , V : Send > Send for IterMut < ' a , K , V > { }
843
855
844
856
/// Iterator over the entries in a table, consuming the table.
845
- pub struct IntoIter < K , V > {
846
- table : RawTable < K , V > ,
847
- iter : RawBuckets < ' static , K , V >
857
+ pub struct IntoIter < K , V , A = HeapAllocator > where A : Allocator {
858
+ table : RawTable < K , V , A > ,
859
+ iter : RawBuckets < ' static , K , V > ,
848
860
}
849
861
850
- unsafe impl < K : Sync , V : Sync > Sync for IntoIter < K , V > { }
851
- unsafe impl < K : Send , V : Send > Send for IntoIter < K , V > { }
862
+ unsafe impl < K : Sync , V : Sync , A : Sync > Sync for IntoIter < K , V , A > where A : Allocator { }
863
+ unsafe impl < K : Send , V : Send , A : Send > Send for IntoIter < K , V , A > where A : Allocator { }
852
864
853
865
/// Iterator over the entries in a table, clearing the table.
854
- pub struct Drain < ' a , K : ' a , V : ' a > {
855
- table : & ' a mut RawTable < K , V > ,
866
+ pub struct Drain < ' a , K : ' a , V : ' a , A : ' a = HeapAllocator > where A : Allocator {
867
+ table : & ' a mut RawTable < K , V , A > ,
856
868
iter : RawBuckets < ' static , K , V > ,
857
869
}
858
870
859
- unsafe impl < ' a , K : Sync , V : Sync > Sync for Drain < ' a , K , V > { }
860
- unsafe impl < ' a , K : Send , V : Send > Send for Drain < ' a , K , V > { }
871
+ unsafe impl < ' a , K : Sync , V : Sync , A : Sync > Sync for Drain < ' a , K , V , A > where A : Allocator { }
872
+ unsafe impl < ' a , K : Send , V : Send , A : Send > Send for Drain < ' a , K , V , A > where A : Allocator { }
861
873
862
874
impl < ' a , K , V > Iterator for Iter < ' a , K , V > {
863
875
type Item = ( & ' a K , & ' a V ) ;
@@ -901,7 +913,7 @@ impl<'a, K, V> ExactSizeIterator for IterMut<'a, K, V> {
901
913
fn len ( & self ) -> usize { self . elems_left }
902
914
}
903
915
904
- impl < K , V > Iterator for IntoIter < K , V > {
916
+ impl < K , V , A > Iterator for IntoIter < K , V , A > where A : Allocator {
905
917
type Item = ( SafeHash , K , V ) ;
906
918
907
919
fn next ( & mut self ) -> Option < ( SafeHash , K , V ) > {
@@ -924,11 +936,11 @@ impl<K, V> Iterator for IntoIter<K, V> {
924
936
( size, Some ( size) )
925
937
}
926
938
}
927
- impl < K , V > ExactSizeIterator for IntoIter < K , V > {
939
+ impl < K , V , A > ExactSizeIterator for IntoIter < K , V , A > where A : Allocator {
928
940
fn len ( & self ) -> usize { self . table . size ( ) }
929
941
}
930
942
931
- impl < ' a , K , V > Iterator for Drain < ' a , K , V > {
943
+ impl < ' a , K , V , A > Iterator for Drain < ' a , K , V , A > where A : Allocator {
932
944
type Item = ( SafeHash , K , V ) ;
933
945
934
946
#[ inline]
@@ -952,20 +964,21 @@ impl<'a, K, V> Iterator for Drain<'a, K, V> {
952
964
( size, Some ( size) )
953
965
}
954
966
}
955
- impl < ' a , K , V > ExactSizeIterator for Drain < ' a , K , V > {
967
+ impl < ' a , K , V , A > ExactSizeIterator for Drain < ' a , K , V , A > where A : Allocator {
956
968
fn len ( & self ) -> usize { self . table . size ( ) }
957
969
}
958
970
959
- impl < ' a , K : ' a , V : ' a > Drop for Drain < ' a , K , V > {
971
+ impl < ' a , K : ' a , V : ' a , A : ' a > Drop for Drain < ' a , K , V , A > where A : Allocator {
960
972
fn drop ( & mut self ) {
961
973
for _ in self { }
962
974
}
963
975
}
964
976
965
- impl < K : Clone , V : Clone > Clone for RawTable < K , V > {
966
- fn clone ( & self ) -> RawTable < K , V > {
977
+ impl < K : Clone , V : Clone , A : Clone > Clone for RawTable < K , V , A > where A : Allocator {
978
+ fn clone ( & self ) -> RawTable < K , V , A > {
967
979
unsafe {
968
- let mut new_ht = RawTable :: new_uninitialized ( self . capacity ( ) ) ;
980
+ let mut new_ht = RawTable :: new_uninitialized_in ( self . capacity ( ) ,
981
+ self . alloc . clone ( ) ) ;
969
982
970
983
{
971
984
let cap = self . capacity ( ) ;
@@ -998,7 +1011,7 @@ impl<K: Clone, V: Clone> Clone for RawTable<K, V> {
998
1011
}
999
1012
}
1000
1013
1001
- impl < K , V > Drop for RawTable < K , V > {
1014
+ impl < K , V , A > Drop for RawTable < K , V , A > where A : Allocator {
1002
1015
#[ unsafe_destructor_blind_to_params]
1003
1016
fn drop ( & mut self ) {
1004
1017
if self . capacity == 0 || self . capacity == mem:: POST_DROP_USIZE {
0 commit comments