当前位置: X-MOL 学术arXiv.cs.AR › 论文详情
Our official English website, www.x-mol.net, welcomes your feedback! (Note: you will need to create a separate account there.)
G-TADOC: Enabling Efficient GPU-Based Text Analytics without Decompression
arXiv - CS - Hardware Architecture Pub Date : 2021-06-13 , DOI: arxiv-2106.06889
Feng Zhang, Zaifeng Pan, Yanliang Zhou, Jidong Zhai, Xipeng Shen, Onur Mutlu, Xiaoyong Du

Text analytics directly on compression (TADOC) has proven to be a promising technology for big data analytics. GPUs are extremely popular accelerators for data analytics systems. Unfortunately, no work so far shows how to utilize GPUs to accelerate TADOC. We describe G-TADOC, the first framework that provides GPU-based text analytics directly on compression, effectively enabling efficient text analytics on GPUs without decompressing the input data. G-TADOC solves three major challenges. First, TADOC involves a large amount of dependencies, which makes it difficult to exploit massive parallelism on a GPU. We develop a novel fine-grained thread-level workload scheduling strategy for GPU threads, which partitions heavily-dependent loads adaptively in a fine-grained manner. Second, in developing G-TADOC, thousands of GPU threads writing to the same result buffer leads to inconsistency while directly using locks and atomic operations lead to large synchronization overheads. We develop a memory pool with thread-safe data structures on GPUs to handle such difficulties. Third, maintaining the sequence information among words is essential for lossless compression. We design a sequence-support strategy, which maintains high GPU parallelism while ensuring sequence information. Our experimental evaluations show that G-TADOC provides 31.1x average speedup compared to state-of-the-art TADOC.

中文翻译:

G-TADOC:无需解压即可实现高效的基于 GPU 的文本分析

直接压缩文本分析 (TADOC) 已被证明是大数据分析的一项有前途的技术。GPU 是非常流行的数据分析系统加速器。不幸的是,到目前为止还没有任何工作展示如何利用 GPU 来加速 TADOC。我们描述了 G-TADOC,这是第一个直接在压缩上提供基于 GPU 的文本分析的框架,有效地在 GPU 上实现高效的文本分析,而无需解压输入数据。G-TADOC 解决了三大挑战。首先,TADOC 涉及大量依赖项,这使得在 GPU 上难以利用大规模并行性。我们为 GPU 线程开发了一种新颖的细粒度线程级工作负载调度策略,它以细粒度的方式自适应地划分高度依赖的负载。其次,在开发 G-TADOC 时,数千个 GPU 线程写入同一个结果缓冲区会导致不一致,而直接使用锁和原子操作会导致大量同步开销。我们在 GPU 上开发了一个具有线程安全数据结构的内存池来处理这些困难。第三,保持单词之间的序列信息对于无损压缩至关重要。我们设计了一个序列支持策略,它在保证序列信息的同时保持高 GPU 并行性。我们的实验评估表明,与最先进的 TADOC 相比,G-TADOC 提供了 31.1 倍的平均加速。保持单词之间的序列信息对于无损压缩至关重要。我们设计了一个序列支持策略,它在保证序列信息的同时保持高 GPU 并行性。我们的实验评估表明,与最先进的 TADOC 相比,G-TADOC 提供了 31.1 倍的平均加速。保持单词之间的序列信息对于无损压缩至关重要。我们设计了一个序列支持策略,它在保证序列信息的同时保持高 GPU 并行性。我们的实验评估表明,与最先进的 TADOC 相比,G-TADOC 提供了 31.1 倍的平均加速。
更新日期:2021-06-15
down
wechat
bug