Key research themes
1. How can communication costs be minimized in parallel LU factorization on large-scale high-performance computing systems?
This research area focuses on deriving theoretical lower bounds for data movement (communication volume) in parallel LU factorization algorithms and designing practical algorithms that approach these bounds. Minimizing communication costs is critical because data movement dominates runtime and energy consumption on distributed-memory and exascale computing systems, where LU factorization is widely used for solving linear systems in scientific computing.
2. What algorithmic and data-structural techniques enable efficient, scalable parallel LU factorization of hierarchical (H-) matrices with dynamic block structures?
Hierarchical matrices (H-matrices) arising from boundary element and partial differential equations provide data-sparse representations enabling efficient approximate factorizations. Parallelizing LU factorization on modern multicore and manycore architectures requires exploiting nested task parallelism with dynamic, non-uniform data structures due to low-rank blocks whose sizes evolve during computation. Addressing this challenge involves advanced task programming models that can manage dependencies despite changing memory layouts while maximizing concurrency and maintaining fine-grained parallel efficiency.
3. How can the relationship between time and energy consumption in multithreaded LU factorization implementations be quantitatively characterized and optimized on multicore processors?
This research theme examines the scalability of LU factorization algorithms in terms of both execution time and energy consumption using multithreading and dynamic voltage and frequency scaling (DVFS) techniques. Understanding these correlations and tradeoffs is essential for optimizing algorithm implementations for energy-efficient high-performance computing, especially as energy constraints become paramount. The goal is to balance performance and power to minimize overall energy use without sacrificing scalability.