Skip to content

Latest commit

 

History

History
165 lines (133 loc) · 10.6 KB

parallel-rustc.md

File metadata and controls

165 lines (133 loc) · 10.6 KB

Parallel Compilation

As of August 2022, the only stage of the compiler that is already parallel is codegen. Some parts of the compiler already have parallel implementations, such as query evaluation, type check and monomorphization, but the general version of the compiler does not include these parallelization functions. To try out the current parallel compiler, one can install rustc from source code with parallel-compiler = true in the config.toml.

The lack of parallelism at other stages (for example, macro expansion) also represents an opportunity for improving compiler performance.

These next few sections describe where and how parallelism is currently used, and the current status of making parallel compilation the default in rustc.

Codegen

During monomorphization the compiler splits up all the code to be generated into smaller chunks called codegen units. These are then generated by independent instances of LLVM running in parallel. At the end, the linker is run to combine all the codegen units together into one binary. This process occurs in the rustc_codegen_ssa::base module.

Data Structures

The underlying thread-safe data-structures used in the parallel compiler can be found in the rustc_data_structures::sync module. These data structures are implemented differently depending on whether parallel-compiler is true.

data structure parallel non-parallel
Lrc std::sync::Arc std::rc::Rc
Weak std::sync::Weak std::rc::Weak
Atomic{Bool}/{Usize}/{U32}/{U64} std::sync::atomic::Atomic{Bool}/{Usize}/{U32}/{U64} (std::cell::Cell<bool/usize/u32/u64>)
OnceCell std::sync::OnceLock std::cell::OnceCell
Lock<T> (parking_lot::Mutex<T>) (std::cell::RefCell)
RwLock<T> (parking_lot::RwLock<T>) (std::cell::RefCell)
MTRef<'a, T> &'a T &'a mut T
MTLock<T> (Lock<T>) (T)
ReadGuard parking_lot::RwLockReadGuard std::cell::Ref
MappedReadGuard parking_lot::MappedRwLockReadGuard std::cell::Ref
WriteGuard parking_lot::RwLockWriteGuard std::cell::RefMut
MappedWriteGuard parking_lot::MappedRwLockWriteGuard std::cell::RefMut
LockGuard parking_lot::MutexGuard std::cell::RefMut
MappedLockGuard parking_lot::MappedMutexGuard std::cell::RefMut
  • These thread-safe data structures interspersed during compilation can cause a lot of lock contention, which actually degrades performance as the number of threads increases beyond 4. This inspires us to audit the use of these data structures, leading to either refactoring to reduce use of shared state, or persistent documentation covering invariants, atomicity, and lock orderings.

  • On the other hand, we still need to figure out what other invariants during compilation might not hold in parallel compilation.

WorkLocal

WorkLocal is a special data structure implemented for parallel compiler. It holds worker-locals values for each thread in a thread pool. You can only access the worker local value through the Deref impl on the thread pool it was constructed on. It will panic otherwise.

WorkLocal is used to implement the Arena allocator in the parallel environment, which is critical in parallel queries. Its implementation is located in the rustc-rayon-core::worker_local module. However, in the non-parallel compiler, it is implemented as (OneThread<T>), whose T can be accessed directly through Deref::deref.

Parallel Iterator

The parallel iterators provided by the rayon crate are easy ways to implement parallelism. In the current implementation of the parallel compiler we use a custom fork of rayon to run tasks in parallel.

Some iterator functions are implemented to run loops in parallel when parallel-compiler is true.

Function(Omit Send and Sync) Introduction Owning Module
par_iter<T: IntoParallelIterator>(t: T) -> T::Iter generate a parallel iterator rustc_data_structure::sync
par_for_each_in<T: IntoParallelIterator>(t: T, for_each: impl Fn(T::Item)) generate a parallel iterator and run for_each on each element rustc_data_structure::sync
Map::par_body_owners(self, f: impl Fn(LocalDefId)) run f on all hir owners in the crate rustc_middle::hir::map
Map::par_for_each_module(self, f: impl Fn(LocalDefId)) run f on all modules and sub modules in the crate rustc_middle::hir::map
ModuleItems::par_items(&self, f: impl Fn(ItemId)) run f on all items in the module rustc_middle::hir
ModuleItems::par_trait_items(&self, f: impl Fn(TraitItemId)) run f on all trait items in the module rustc_middle::hir
ModuleItems::par_impl_items(&self, f: impl Fn(ImplItemId)) run f on all impl items in the module rustc_middle::hir
ModuleItems::par_foreign_items(&self, f: impl Fn(ForeignItemId)) run f on all foreign items in the module rustc_middle::hir

There are a lot of loops in the compiler which can possibly be parallelized using these functions. As of August 2022, scenarios where the parallel iterator function has been used are as follows:

caller scenario callee
rustc_metadata::rmeta::encoder::prefetch_mir Prefetch queries which will be needed later by metadata encoding par_iter
rustc_monomorphize::collector::collect_crate_mono_items Collect monomorphized items reachable from non-generic items par_for_each_in
rustc_interface::passes::analysis Check the validity of the match statements Map::par_body_owners
rustc_interface::passes::analysis MIR borrow check Map::par_body_owners
rustc_typeck::check::typeck_item_bodies Type check Map::par_body_owners
rustc_interface::passes::hir_id_validator::check_crate Check the validity of hir Map::par_for_each_module
rustc_interface::passes::analysis Check the validity of loops body, attributes, naked functions, unstable abi, const bodys Map::par_for_each_module
rustc_interface::passes::analysis Liveness and intrinsic checking of MIR Map::par_for_each_module
rustc_interface::passes::analysis Deathness checking Map::par_for_each_module
rustc_interface::passes::analysis Privacy checking Map::par_for_each_module
rustc_lint::late::check_crate Run per-module lints Map::par_for_each_module
rustc_typeck::check_crate Well-formedness checking Map::par_for_each_module

There are still many loops that have the potential to use parallel iterators.

Query System

The query model has some properties that make it actually feasible to evaluate multiple queries in parallel without too much of an effort:

  • All data a query provider can access is accessed via the query context, so the query context can take care of synchronizing access.
  • Query results are required to be immutable so they can safely be used by different threads concurrently.

When a query foo is evaluated, the cache table for foo is locked.

  • If there already is a result, we can clone it, release the lock and we are done.
  • If there is no cache entry and no other active query invocation computing the same result, we mark the key as being "in progress", release the lock and start evaluating.
  • If there is another query invocation for the same key in progress, we release the lock, and just block the thread until the other invocation has computed the result we are waiting for. Cycle error detection in the parallel compiler requires more complex logic than in single-threaded mode. When worker threads in parallel queries stop making progress due to interdependence, the compiler uses an extra thread (named deadlock handler) to detect, remove and report the cycle error.

Parallel query still has a lot of work to do, most of which is related to the previous Data Structures and Parallel Iterators. See this tracking issue.

Rustdoc

As of November 2022, there are still a number of steps to complete before rustdoc rendering can be made parallel. More details on this issue can be found here.

Resources

Here are some resources that can be used to learn more (note that some of them are a bit out of date):