Accelerating hash-based query processing operations on FPGAs by a hash table caching technique

4Citations
Citations of this article
7Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Extracting valuable information from the rapidly growing field of Big Data faces serious performance constraints, especially in the software-based database management systems (DBMS). In a query processing system, hash-based computational primitives such as the hash join and the group-by are the most time-consuming operations, as they frequently need to access the hash table on the high-latency off-chip memories and also to traverse whole the table. Subsequently, the hash collision is an inherent issue related to the hash tables, which can adversely degrade the overall performance. In order to alleviate this problem, in this paper, we present a novel pure hardware-based hash engine, implemented on the FPGA. In order to mitigate the high memory access latencies and also to faster resolve the hash collisions, we follow a novel design point. It is based on caching the hash table entries in the fast on-chip Block-RAMs of FPGA. Faster accesses to the correspondent hash table entries from the cache can lead to an improved overall performance. We evaluate the proposed approach by running hash-based table join and group-by operations of 5 TPC-H benchmark queries. The results show 2.9×– 4.4× speedups over the cache-less FPGA-based baseline.

Cite

CITATION STYLE

APA

Salami, B., Arcas-Abella, O., Sonmez, N., Unsal, O., & Kestelman, A. C. (2017). Accelerating hash-based query processing operations on FPGAs by a hash table caching technique. In Communications in Computer and Information Science (Vol. 697, pp. 131–145). Springer Verlag. https://doi.org/10.1007/978-3-319-57972-6_10

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free