Results 51 to 60 of about 42,875 (250)
An Architecture for a Tri-Programming Model-Based Parallel Hybrid Testing Tool
As the development of high-performance computing (HPC) is growing, exascale computing is on the horizon. Therefore, it is imperative to develop parallel systems, such as graphics processing units (GPUs) and programming models, that can effectively ...
Saeed Musaad Altalhi +6 more
doaj +1 more source
StePS: A Multi-GPU Cosmological N-body Code for Compactified Simulations [PDF]
We present the multi-GPU realization of the StePS (Stereographically Projected Cosmological Simulations) algorithm with MPI-OpenMP-CUDA hybrid parallelization and nearly ideal scale-out to multiple compute nodes.
Csabai, István +4 more
core +2 more sources
В статье представлено исследование вычислительной эффективности гибридного параллельного алгоритма, реализующего нелокальную неявную конечно-разностную схему (IFDS) для численного решения задачи динамики объемной активности радона (ОАР).
Твёрдый, Д.А.
doaj +1 more source
Parallel Simulations of Dynamic Earthquake Rupture along Geometrically Complex Faults on CMP Systems
Chip multiprocessors (CMP) are widely used for high performance computing and are being configured in a hierarchical manner to compose a CMP compute node in a CMP system.
Xingfu Wu, Benchun Duan, Valerie Taylor
doaj +1 more source
Characterizing task-based OpenMP programs. [PDF]
Programmers struggle to understand performance of task-based OpenMP programs since profiling tools only report thread-based performance. Performance tuning also requires task-based performance in order to balance per-task memory hierarchy utilization ...
Ananya Muddukrishna +2 more
doaj +1 more source
Implementing implicit OpenMP data sharing on GPUs
OpenMP is a shared memory programming model which supports the offloading of target regions to accelerators such as NVIDIA GPUs. The implementation in Clang/LLVM aims to deliver a generic GPU compilation toolchain that supports both the native CUDA C/C++
Bataev, Alexey +8 more
core +1 more source
Heterogeneous computing is increasingly being used in a diversity of computing systems, ranging from HPC to the real-time embedded domain, to cope with the performance requirements.
Chen Yu, Sara Royuela, Eduardo Quiñones
semanticscholar +1 more source
To extract performance from supercomputers, programmers in the High Performance Computing (HPC) community are often required to use a combination of frameworks to take advantage of the multiple levels of parallelism. However, over the years, efforts have
Christian Asch +2 more
doaj +1 more source
CoreTSAR: Task Scheduling for Accelerator-aware Runtimes [PDF]
Heterogeneous supercomputers that incorporate computational accelerators such as GPUs are increasingly popular due to their high peak performance, energy efficiency and comparatively low cost.
de Supinski, Bronis R. +3 more
core +1 more source
Peachy Parallel Assignments (EduHPC 2018) [PDF]
Peachy Parallel Assignments are a resource for instructors teaching parallel and distributed programming. These are high-quality assignments, previously tested in class, that are readily adoptable.
Alvarez, Lluc +10 more
core +2 more sources

