-
Notifications
You must be signed in to change notification settings - Fork 0
Home
Highly controversial, and possibly not much beneficial, attempt to improve (CPU) cache efficiency (and, hopefully, but less likely, decrease CPU cycles) for heap (de)allocation/resizing in Rust. Specific/only applicable to Rust (or languages/methodologies with RAII and with support for custom allocators).
Not related/specific to Rust co-routines (async/await
), neither to any cooperative multithreading/multitasking/scheduling.
Leverage Rust memory & data safety guarantees to streamline de-allocation (and resizing of data on heap).
When de-allocating, all that Rust passes through the (standard) allocator API is a pointer (to the data being de-allocated). The allocator has to look it up for its housekeeping in its internal data structures (often hash-based).
Rust could keep some extra metadata along every smart pointer. (This metadata would come from the allocator's new functions co_allocate, co_allocate_zeroed, co_grow, co_grow_zeroed, co_shrink
.) Then Rust could pass that metadata back (to the allocator) when de-allocating (or resizing). That hopefully saves the allocator (listing the highest benefit first):
- data cache access (looking up auxiliary/hashmap-like record(s) for the given pointer, or its hash, and/or the allocated data length...), and
- code cache access (no need to re-calculate the hash, or its precursor), and
- CPU cycles (calculating the hash, or its precursor, and locating the related entries) - least likely benefit (since this gain is decreased by CPU cycles needed to move the metadata around).
The size of the metadata (per smart pointer) must be known in compile time. Hence, this is supported only for the global
allocator, which must define the metadata and make it (somehow) known to the compiler. That may need some new compiler configuration/directives and/or rustc
target configuration.
However, this doesn't affect Rust language itself (other than fixing some rustc
internal compiler errors). (Would you like this to be clearer? Get in touch & help shape this, or at least shine the light, please.)
Of course, this increases the size of (affected) smart pointers (wherever used/stored/passed). So, the initial candidates are primarily non-leaf smart pointers (Vec
and other containers):
- Passed around (up & down the stack, or stored/moved around on the heap), in total, less than leaf objects. (As data flow accumulated over time. Well designed data structures have more leaves than non-leaves for common input.)
- Small in total size (as per the previous). Then any small metadata (a small multiple of
usize
, adjacent to those smart pointers) may not increase RAM requirements too much. - So, (as per the previous two) this metadata may not excessively affect stack depth, neither the data flow & CPU cycles (for many applications).
- Due to the cache waiting times, handling the metadata may increase execution time even less so than it increases the number of CPU cycles (since the CPU waits for the cache anyway). Especially so when these smart pointers are stored on stack (which grows/shrinks sequentially). Therefore the runtime speed cost of handling the metadata may not be as high as it may sound.
- Resizable (which can make use of the metadata):
Vec
(and hopefullyHashMap/HashSet, BTreeMap/BTreeSet
). - (De)allocated or resized frequently. Hence
String
, or any frequently resizedVec
of primitive orstatic
data, too (even though it's a leaf in terms of smart pointer "levels".) - Of course, leaf smart pointers are more frequently (de)allocated than non-leaf ones (in well designed data structures). But the overall data flow may prohibit them (from using cooperative (de)allocation - unless they are resizable).
Hence, (in general) this would be much less beneficial (or even detrimental) for Box
, Rc
, Arc
. (Or, if anyone implements this for those smart pointers, too, we would see when we are benchmarking this.)
Allocators can store metadata (together with the allocated data) - even without passing it around. How? They can put it "left" of the allocated data: At (consecutive) addresses just below the returned pointer. Then the allocator could read such metadata (and update it, if necessary) on deallocation/resize. This is allocator-agnostic, as the size of metadata can vary from one allocator to another.
However, such allocation, deallocation and resize has to access that metadata from that location (next to the user's allocated data). If the application is not using the allocated data near that moment, this increases the cache reloads.
Also, many resizable data structures (like Vec
) are designed to grow in multiples of cache page length. If this (adjacent) metadata size varies between allocators, the data structure size may be more difficult to optimize (to make the total size a multiple of cache page length).
TODO Combine both: Passing metadata around, or storing it left of the allocated data (or neither). Have this configurable: Instead of a bool
const generic, pass an enum
representing this preference. Then have two where
bounds:
- one for
usize
of metadata passed around, and - one for the
usize
of metadata stored left of the allocated data.
Then change the initial size: Subtract the size of metadata left of the allocated data (so that it all fits into a multiple of cache page length).
Whether a type is "cooperative" is highly configurable with const
generics. The user (developer) provides value of a const
generic bool
parameter called COOP_PREFERRED
. That indicates whether this type should use cooperative (de)allocation.
There are also "weighted" versions (for Vec
such version is WeVec
- name can be discussed). Such types accept an integral const
generic "weight" (suggested u8
). Then the actual rustc
target, and/or some configuration parameter (to be discussed), specify the threshold. "Weighted" types with their "weight" above the threshold operate as "cooperative"; otherwise they operate as "non-cooperative".
If the user doesn't provide a const
generic parameter COOP_PREFERRED
, then the type uses its default. Again, that will (hopefully) be configurable per build (or at least per target).
Adding the new functions to the Allocator
API (listed above; with defaults/fallbacks) is easy. Implementing them is up to the allocator, and out of scope here. (Do you know of any allocators written primarily in Rust
- rather than in C - with developers open to adding new functionality? Please connect us.)
However, most of the work in Rust (alloc
/std
and core
) is about storing and handling the metadata, and adding the const
generic parameters (and their where
bounds)for the developer's choices and defaults.
This affects a lot of Vec, VecDeque
(and RawVec) in
library/alloc(which is aliased to
library/std). Implementing this for
String affects
Cowand
ToOwned(in
library/core`), too.
Questions
- Names of new methods (
Vec::new_co
) and types (`WeVec - see above)... - There are smart pointers (and similar/related) structs/enums/traits that currently only work with the
global
allocator (for example,String
). When adding theCOOP_PREFERRED
preference flag, should we add the allocator generic parameter, too? (Then we'd need related methods that accept the allocator instance, similar to the existingVec::new_in
.) If we ever want this, it has to be now.
Types with different COOP_PREFERRED
are not compatible.
"Downgrading" from a "cooperative" type to its "non-cooperative" version is zero cost.
However, "upgrading" is not possible. That's unless we make the "cooperative" version always check its metadata to be non-null first. (That would have a fallback to non-cooperative allocator API.)
HIGHLY safe. TODO explain here.
Safe Rust code will work fine. unsafe
code will work, even if it dismantles raw parts of Vec
(and similar), as far as it uses the type's functions (rather than using unsafe
to cast the smart pointer parts to some other types and extract parts manually).
However, TODO Fix/check MVP code where constructing Vec
and similar from raw parts. Investigate if we could make it work for non-cooperative versions only. (E.g. panic!
if called to construct a cooperative version.)
Existing crates will stay compatible.
HIGHLY compatible with const
generic default values. If a crate doesn't specify COOP_PREFERRED
, it will work with whatever COOP_PREFERRED
the consumer crate chooses.
TODO explain more.
- Primarily, and initially,
Vec
,VecDeque
andRawVec
(MVP complete, not tested, not benchmarked). - Possibly
String
and related (Cow
,ToOwned
). (MVP mostly complete, not tested, not benchmarked. Please help.) - Hopefully
HashMap
/HashSet
,BTreeMap
/BTreeSet
) - Unlikely (possible, but with little benefit - as per above), and more involving the compiler & the language internals:
Box
,Rc
,Arc
.
See Contributing.