skip to main content
Volume 11, Issue 2June 2024Current Issue
Bibliometrics
Skip Table Of Content Section
research-article
TLPGNN: A Lightweight Two-level Parallelism Paradigm for Graph Neural Network Computation on Single and Multiple GPUs
Article No.: 7, Pages 1–28https://doi.org/10.1145/3644712

Graph Neural Networks (GNNs) are an emerging class of deep learning models specifically designed for graph-structured data. They have been effectively employed in a variety of real-world applications, including recommendation systems, drug development, ...

research-article
cuFasterTucker: A Stochastic Optimization Strategy for Parallel Sparse FastTucker Decomposition on GPU Platform
Article No.: 8, Pages 1–33https://doi.org/10.1145/3648094

The amount of scientific data is currently growing at an unprecedented pace, with tensors being a common form of data that display high-order, high-dimensional, and sparse features. While tensor-based analysis methods are effective, the vast increase in ...

research-article
Low-Overhead Trace Collection and Profiling on GPU Compute Kernels
Article No.: 9, Pages 1–24https://doi.org/10.1145/3649510

While GPUs can bring substantial speedup to compute-intensive tasks, their programming is notoriously hard. From their programming model, to microarchitectural particularities, the programmer may encounter many pitfalls which may hinder performance in ...

research-article
Decentralized Scheduling for Data-Parallel Tasks in the Cloud
Article No.: 10, Pages 1–23https://doi.org/10.1145/3651858

For latency-sensitive data processing applications in the cloud, concurrent data-parallel tasks need to be scheduled and processed quickly. A data-parallel task usually consists of a set of sub-tasks, generating a set of flows that are collectively ...

research-article
Machine Learning-Based Kernel Selector for SpMV Optimization in Graph Analysis
Article No.: 11, Pages 1–25https://doi.org/10.1145/3652579

Sparse Matrix and Vector multiplication (SpMV) is one of the core algorithms in various large-scale scientific computing and real-world applications. With the rapid development of AI and big data, the input vector in SpMV becomes sparse in many ...

research-article
cuFastTucker: A Novel Sparse FastTucker Decomposition For HHLST on Multi-GPUs
Article No.: 12, Pages 1–31https://doi.org/10.1145/3661450

High-order, high-dimension, and large-scale sparse tensors (HHLST) have found their origin in various real industrial applications, such as social networks, recommender systems, bioinformatics, and traffic information. To handle these complex tensors, ...

Subjects

Comments