allocative/size_of.rs
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192
/*
* Copyright (c) Meta Platforms, Inc. and affiliates.
*
* This source code is licensed under both the MIT license found in the
* LICENSE-MIT file in the root directory of this source tree and the Apache
* License, Version 2.0 found in the LICENSE-APACHE file in the root directory
* of this source tree.
*/
use crate::visitor::NodeKind;
use crate::visitor::VisitorImpl;
use crate::Allocative;
use crate::Key;
use crate::Visitor;
/// Size of data allocated in unique pointers in the struct.
///
/// * Exclude self
/// * Exclude shared pointers
/// * For unique pointers, include the size of the pointee plus this function recursively
///
/// # Example
///
/// ```
/// use allocative::Allocative;
///
/// #[derive(Allocative)]
/// struct Foo {
/// data: Vec<u8>,
/// }
///
/// assert_eq!(
/// 3,
/// allocative::size_of_unique_allocated_data(&Foo {
/// data: vec![10, 20, 30]
/// })
/// );
/// ```
pub fn size_of_unique_allocated_data(root: &dyn Allocative) -> usize {
struct SizeOfUniqueAllocatedDataVisitor {
/// Size we return.
size: usize,
}
impl VisitorImpl for SizeOfUniqueAllocatedDataVisitor {
fn enter_inline_impl(&mut self, _name: Key, size: usize, parent: NodeKind) {
if let NodeKind::Unique = parent {
self.size += size;
}
}
fn enter_unique_impl(&mut self, _name: Key, _size: usize, _parent: NodeKind) {}
fn enter_shared_impl(
&mut self,
_name: Key,
_size: usize,
_ptr: *const (),
_parent: NodeKind,
) -> bool {
false
}
fn exit_inline_impl(&mut self) {}
fn exit_unique_impl(&mut self) {}
fn exit_shared_impl(&mut self) {
unreachable!("shared pointers are not visited")
}
fn exit_root_impl(&mut self) {}
}
let mut visitor_impl = SizeOfUniqueAllocatedDataVisitor { size: 0 };
let mut visitor = Visitor {
visitor: &mut visitor_impl,
node_kind: NodeKind::Root,
};
root.visit(&mut visitor);
visitor.exit();
visitor_impl.size
}
/// Size of a piece of data and data allocated in unique pointers in the struct.
///
/// * Excludes shared pointers
///
/// # Example
///
/// ```
/// use allocative::Allocative;
///
/// #[derive(Allocative)]
/// struct Foo {
/// data: Vec<u8>,
/// }
///
/// assert_eq!(
/// 3 + std::mem::size_of::<Vec<u8>>(),
/// allocative::size_of_unique(&Foo {
/// data: vec![10, 20, 30]
/// })
/// );
/// ```
pub fn size_of_unique<T>(root: &T) -> usize
where
T: Allocative,
{
std::mem::size_of::<T>() + size_of_unique_allocated_data(root)
}
#[cfg(test)]
mod tests {
use std::mem;
use allocative_derive::Allocative;
use crate as allocative;
use crate::size_of_unique;
use crate::size_of_unique_allocated_data;
#[test]
fn test_box() {
#[derive(Allocative)]
struct Boxed {
data: Box<u32>,
}
let boxed = Boxed { data: Box::new(17) };
assert_eq!(mem::size_of::<u32>(), size_of_unique_allocated_data(&boxed));
assert_eq!(
mem::size_of::<u32>() + mem::size_of::<Boxed>(),
size_of_unique(&boxed)
);
}
#[test]
fn test_box_slice() {
#[derive(Allocative)]
struct Boxed {
data: Box<[u32]>,
}
let boxed = Boxed {
data: vec![1, 2, 3].into_boxed_slice(),
};
assert_eq!(
mem::size_of::<u32>() * 3,
size_of_unique_allocated_data(&boxed)
);
assert_eq!(
mem::size_of::<Boxed>() + mem::size_of::<u32>() * 3,
size_of_unique(&boxed)
);
}
#[test]
fn test_struct_in_box() {
#[derive(Allocative)]
struct Data {
a: u8,
b: Box<u32>,
}
#[derive(Allocative)]
struct Boxed {
data: Box<Data>,
}
let boxed = Boxed {
data: Box::new(Data {
a: 1,
b: Box::new(2),
}),
};
assert_eq!(
mem::size_of::<Data>() + mem::size_of::<u32>(),
size_of_unique_allocated_data(&boxed)
);
assert_eq!(
mem::size_of::<Boxed>() + mem::size_of::<Data>() + mem::size_of::<u32>(),
size_of_unique(&boxed)
);
}
}