Auto-tuning of dynamic load balancing applied to 3D reverse time migration on multicore systems

05/16/2019
by   Ítalo A. S. Assis, et al.
0

Reverse time migration (RTM) is an algorithm widely used in the oil and gas industry to process seismic data. It is a computationally intensive task that suits well in parallel computers. Because of it being massive and regular, this type of task is often equally and statically distributed among the available parallel workers. However, this strategy is often not optimal. When the workers are heterogeneous, and even when most have similar processing power, many of them might still have to wait idly for the slower workers. In this paper, we show that even small performance differences between homogeneous cores can considerably affect the overall performance of a 3D RTM application. We show that dynamic load distribution has a significant advantage over the conventional static distribution. However, the granularity of the dynamically distributed chunks of work plays a key role in harvesting this advantage. In order to find the optimal granularity, we propose a coupled simulated annealing (CSA) based auto-tuning strategy that adjusts the chunk size of work that OpenMP parallel loops assign dynamically to worker threads during the initialization of a 3D RTM application. Experiments performed on computational systems with different processor and RAM specifications and for different sizes of input show that the proposed method is consistently better than two default OpenMP loop schedulers being up to 44

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/01/2019

Distributed-Memory Load Balancing with Cyclic Token-based Work-Stealing Applied to Reverse Time Migration

Reverse time migration (RTM) is a prominent technique in seismic imaging...
research
01/14/2018

Distributed dynamic load balancing for task parallel programming

In this paper, we derive and investigate approaches to dynamically load ...
research
10/03/2021

Distributed Optimization using Heterogeneous Compute Systems

Hardware compute power has been growing at an unprecedented rate in rece...
research
02/06/2019

Scheduling and Trade-off Analysis for Multi-Source Multi-Processor Systems with Divisible Loads

The main goal of parallel processing is to provide users with performanc...
research
10/01/2018

Heterogeneous MacroTasking (HeMT) for Parallel Processing in the Public Cloud

Using tiny, equal-sized tasks (Homogeneous microTasking, HomT) has long ...
research
11/30/2021

Atos: A Task-Parallel GPU Dynamic Scheduling Framework for Dynamic Irregular Computations

We present Atos, a task-parallel GPU dynamic scheduling framework that i...
research
07/01/2018

Ensuring domain consistency in an adaptive framework with distributed topology for fluid flow simulations

Top-tier parallel computing clusters continue to accumulate more and mor...

Please sign up or login with your details

Forgot password? Click here to reset