Skip to content

Commit 1e561b9

Browse files
committed
Add .shrink_to_fit() for Array
Add `.shrink_to_fit()` for owned storage arrays. This initial version does not change array strides, this is to avoid the most tricky part of the implementation (and maybe take it step by step). Care must be taken since `self.ptr` points inside the allocation - which for `Array` is still `Vec`-allocated, and several Vec methods including its `shrink_to_fit` can reallocate the array.
1 parent 6f77377 commit 1e561b9

File tree

4 files changed

+211
-1
lines changed

4 files changed

+211
-1
lines changed

src/data_repr.rs

+42
Original file line numberDiff line numberDiff line change
@@ -4,6 +4,7 @@ use alloc::borrow::ToOwned;
44
use alloc::slice;
55
#[cfg(not(feature = "std"))]
66
use alloc::vec::Vec;
7+
use core::ops::Range;
78
use std::mem;
89
use std::mem::ManuallyDrop;
910
use std::ptr::NonNull;
@@ -28,6 +29,9 @@ pub struct OwnedRepr<A>
2829
capacity: usize,
2930
}
3031

32+
// OwnedRepr is a wrapper for a uniquely held allocation. Currently it is allocated by using a Vec
33+
// (from/to raw parts) which gives the benefit that it can always be converted to/from a Vec
34+
// cheaply.
3135
impl<A> OwnedRepr<A>
3236
{
3337
pub(crate) fn from(v: Vec<A>) -> Self
@@ -54,6 +58,14 @@ impl<A> OwnedRepr<A>
5458
self.len
5559
}
5660

61+
#[cfg(test)]
62+
/// Note: Capacity comes from OwnedRepr (Vec)'s allocation strategy and cannot be absolutely
63+
/// guaranteed.
64+
pub(crate) fn capacity(&self) -> usize
65+
{
66+
self.capacity
67+
}
68+
5769
pub(crate) fn as_ptr(&self) -> *const A
5870
{
5971
self.ptr.as_ptr()
@@ -85,6 +97,36 @@ impl<A> OwnedRepr<A>
8597
self.as_nonnull_mut()
8698
}
8799

100+
/// Return the new lowest address pointer of the allocation.
101+
#[must_use = "must use new pointer to update existing pointers"]
102+
pub(crate) fn shrink_to_fit(&mut self) -> NonNull<A>
103+
{
104+
// Vec::shrink_to_fit is allowed to reallocate and invalidate pointers
105+
self.modify_as_vec(|mut v| {
106+
v.shrink_to_fit();
107+
v
108+
});
109+
self.as_nonnull_mut()
110+
}
111+
112+
/// Truncate "at front and back", preserve only elements inside the range,
113+
/// then call shrink_to_fit().
114+
/// Moving elements will invalidate existing pointers.
115+
///
116+
/// Return the new lowest address pointer of the allocation.
117+
#[must_use = "must use new pointer to update existing pointers"]
118+
pub(crate) fn preserve_range_and_shrink(&mut self, span: Range<usize>) -> NonNull<A>
119+
{
120+
self.modify_as_vec(|mut v| {
121+
v.truncate(span.end);
122+
if span.start > 0 {
123+
v.drain(..span.start);
124+
}
125+
v
126+
});
127+
self.shrink_to_fit()
128+
}
129+
88130
/// Set the valid length of the data
89131
///
90132
/// ## Safety

src/dimension/mod.rs

+17-1
Original file line numberDiff line numberDiff line change
@@ -428,7 +428,7 @@ fn to_abs_slice(axis_len: usize, slice: Slice) -> (usize, usize, isize)
428428

429429
/// This function computes the offset from the lowest address element to the
430430
/// logically first element.
431-
pub fn offset_from_low_addr_ptr_to_logical_ptr<D: Dimension>(dim: &D, strides: &D) -> usize
431+
pub(crate) fn offset_from_low_addr_ptr_to_logical_ptr<D: Dimension>(dim: &D, strides: &D) -> usize
432432
{
433433
let offset = izip!(dim.slice(), strides.slice()).fold(0, |_offset, (&d, &s)| {
434434
let s = s as isize;
@@ -442,6 +442,22 @@ pub fn offset_from_low_addr_ptr_to_logical_ptr<D: Dimension>(dim: &D, strides: &
442442
offset as usize
443443
}
444444

445+
/// This function computes the offset from the logically first element to the highest address
446+
/// element.
447+
pub(crate) fn offset_from_logical_ptr_to_high_addr_ptr<D: Dimension>(dim: &D, strides: &D) -> usize
448+
{
449+
let offset = izip!(dim.slice(), strides.slice()).fold(0, |_offset, (&d, &s)| {
450+
let s = s as isize;
451+
if s > 0 && d > 1 {
452+
_offset + s * (d as isize - 1)
453+
} else {
454+
_offset
455+
}
456+
});
457+
debug_assert!(offset >= 0);
458+
offset as usize
459+
}
460+
445461
/// Modify dimension, stride and return data pointer offset
446462
///
447463
/// **Panics** if stride is 0 or if any index is out of bounds.

src/impl_owned_array.rs

+118
Original file line numberDiff line numberDiff line change
@@ -859,6 +859,57 @@ where D: Dimension
859859

860860
Ok(())
861861
}
862+
863+
/// Shrink Array allocation capacity to be as small as it can be.
864+
pub fn shrink_to_fit(&mut self)
865+
{
866+
// Example:
867+
// (1) (2) (3) .- len
868+
// Vector: [ x x x x x V x V x V x V x V x V x V x V x V x x x x x x x ] .- capacity
869+
// Allocation: [ m m m m m m m m m m m m m m m m m m m m m m m m m m m m m m m m m ]
870+
//
871+
// x: valid data in OwnedRepr but outside current array slicing
872+
// V: valid data in OwnedRepr and visible in current array slicing
873+
// m: allocated memory
874+
// (1): Lowest address element
875+
// (2): Logical pointer (Element at index zero; normally (1) == (2) but can be
876+
// located anywhere (1) <= (2) <= (3))
877+
// (3): Highest address element
878+
//
879+
// Span: From (1) to (3).
880+
//
881+
// Algorithm: Compute 1, 2, 3.
882+
// Move data so that unused areas before (1) and after (3) are removed from the storage/vector.
883+
// Then shrink the vector's allocation to fit the valid elements.
884+
//
885+
// After:
886+
// (1) (2) (3).- len == capacity
887+
// Vector: [ V x V x V x V x V x V x V x V x V ]
888+
// Allocation: [ m m m m m m m m m m m m m m m m m ]
889+
//
890+
891+
if mem::size_of::<A>() == 0 {
892+
return;
893+
}
894+
895+
let data_ptr = self.data.as_ptr();
896+
let logical_ptr = self.as_ptr();
897+
let offset_to_logical = dimension::offset_from_low_addr_ptr_to_logical_ptr(&self.dim, &self.strides);
898+
let offset_to_high = dimension::offset_from_logical_ptr_to_high_addr_ptr(&self.dim, &self.strides);
899+
900+
let span = offset_to_logical + offset_to_high + 1;
901+
debug_assert!(span >= self.len());
902+
903+
let guard = AbortIfPanic(&"shrink_to_fit: owned repr not in consistent state");
904+
unsafe {
905+
let front_slop = logical_ptr.offset_from(data_ptr) as usize - offset_to_logical;
906+
let new_low_ptr = self
907+
.data
908+
.preserve_range_and_shrink(front_slop..(front_slop + span));
909+
self.ptr = new_low_ptr.add(offset_to_logical);
910+
}
911+
guard.defuse();
912+
}
862913
}
863914

864915
/// This drops all "unreachable" elements in `self_` given the data pointer and data length.
@@ -1016,3 +1067,70 @@ where D: Dimension
10161067
}
10171068
}
10181069
}
1070+
1071+
#[cfg(test)]
1072+
mod tests
1073+
{
1074+
use crate::Array;
1075+
use crate::Array2;
1076+
use crate::Slice;
1077+
use core::fmt::Debug;
1078+
use core::mem::size_of;
1079+
1080+
#[test]
1081+
fn test_shrink_to_fit()
1082+
{
1083+
fn assert_shrink_before_after<T>(mut a: Array2<T>, s1: Slice, s2: Slice, new_capacity: usize)
1084+
where T: Debug + Clone + Eq
1085+
{
1086+
let initial_len = a.len();
1087+
if size_of::<T>() > 0 {
1088+
assert_eq!(a.data.capacity(), initial_len);
1089+
}
1090+
a = a.slice_move(s![s1, s2]);
1091+
let before_value = a.clone();
1092+
let before_strides = a.strides().to_vec();
1093+
#[cfg(feature = "std")]
1094+
println!("{:?}, {}, {:?}", a, a.len(), a.data);
1095+
a.shrink_to_fit();
1096+
#[cfg(feature = "std")]
1097+
println!("{:?}, {}, {:?}", a, a.len(), a.data);
1098+
1099+
assert_eq!(before_value, a);
1100+
assert_eq!(before_strides, a.strides());
1101+
1102+
if size_of::<T>() > 0 {
1103+
assert!(a.data.capacity() < initial_len);
1104+
assert!(a.data.capacity() >= a.len());
1105+
}
1106+
assert_eq!(a.data.capacity(), new_capacity);
1107+
}
1108+
1109+
let a = Array::from_iter(0..56)
1110+
.into_shape_with_order((8, 7))
1111+
.unwrap();
1112+
assert_shrink_before_after(a, Slice::new(1, Some(-1), 1), Slice::new(0, None, 2), 42);
1113+
1114+
let a = Array::from_iter(0..56)
1115+
.into_shape_with_order((8, 7))
1116+
.unwrap();
1117+
assert_shrink_before_after(a, Slice::new(1, Some(-1), -1), Slice::new(0, None, -1), 42);
1118+
1119+
let a = Array::from_iter(0..56)
1120+
.into_shape_with_order((8, 7))
1121+
.unwrap();
1122+
assert_shrink_before_after(a, Slice::new(1, Some(3), 1), Slice::new(1, None, -2), 12);
1123+
1124+
// empty but still has some allocation to allow offsetting along each stride
1125+
let a = Array::from_iter(0..56)
1126+
.into_shape_with_order((8, 7))
1127+
.unwrap();
1128+
assert_shrink_before_after(a, Slice::new(1, Some(1), 1), Slice::new(1, None, 1), 6);
1129+
1130+
// Test ZST
1131+
let a = Array::from_iter((0..56).map(|_| ()))
1132+
.into_shape_with_order((8, 7))
1133+
.unwrap();
1134+
assert_shrink_before_after(a, Slice::new(1, Some(3), 1), Slice::new(1, None, -2), usize::MAX);
1135+
}
1136+
}

tests/assign.rs

+34
Original file line numberDiff line numberDiff line change
@@ -232,6 +232,39 @@ fn move_into()
232232
}
233233
}
234234

235+
#[test]
236+
fn shrink_to_fit_slicing()
237+
{
238+
// Count correct number of drops when using shrink_to_fit and discontiguous arrays (with holes).
239+
for &use_f_order in &[false, true] {
240+
for &invert_axis in &[0b00, 0b01, 0b10, 0b11] {
241+
// bitmask for axis to invert
242+
let counter = DropCounter::default();
243+
{
244+
let (m, n) = (5, 4);
245+
246+
let mut a = Array::from_shape_fn((m, n).set_f(use_f_order), |_idx| counter.element());
247+
a.slice_collapse(s![1..-1, ..;2]);
248+
if invert_axis & 0b01 != 0 {
249+
a.invert_axis(Axis(0));
250+
}
251+
if invert_axis & 0b10 != 0 {
252+
a.invert_axis(Axis(1));
253+
}
254+
255+
a.shrink_to_fit();
256+
257+
let total = m * n;
258+
let dropped_1 = if use_f_order { n * 2 - 1 } else { m * 2 - 1 };
259+
assert_eq!(counter.created(), total);
260+
assert_eq!(counter.dropped(), dropped_1 as usize);
261+
drop(a);
262+
}
263+
counter.assert_drop_count();
264+
}
265+
}
266+
}
267+
235268
/// This counter can create elements, and then count and verify
236269
/// the number of which have actually been dropped again.
237270
#[derive(Default)]
@@ -241,6 +274,7 @@ struct DropCounter
241274
dropped: AtomicUsize,
242275
}
243276

277+
#[derive(Debug)]
244278
struct Element<'a>(&'a AtomicUsize);
245279

246280
impl DropCounter

0 commit comments

Comments
 (0)