Light loss-less data compression, with GPU implementation

10Citations
Citations of this article
5Readers
Mendeley users who have this article in their library.
Get full text

Abstract

There is no doubt that data compression is very important in computer engineering. However, most lossless data compression and decompression algorithms are very hard to parallelize, because they use dictionaries updated sequentially. The main contribution of this paper is to present a new lossless data compression method that we call Light Loss-Less (LLL) compression. It is designed so that decompression can be highly parallelized and run very efficiently on the GPU. This makes sense for many applications in which compressed data is read and decompressed many times and decompression performed more frequently than compression. We show optimal sequential and parallel algorithms for LLL decompression and implement them to run on Core i7-4790 CPU and GeForce GTX 1080 GPU, respectively. To show the potentiality of LLL compression method, we have evaluated the running time using five images and compared with well-known compression methods LZW and LZSS. Our GPU implementation of LLL decompression runs 91.1–176 times faster than the CPU implementation. Also, the running time on the GPU of our experiments show that LLL decompression is 2.49–9.13 times faster than LZW decompression and 4.30–14.1 times faster that LZSS decompression, although their compression ratios are comparable.

Cite

CITATION STYLE

APA

Funasaka, S., Nakano, K., & Ito, Y. (2016). Light loss-less data compression, with GPU implementation. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 10048 LNCS, pp. 281–294). Springer Verlag. https://doi.org/10.1007/978-3-319-49583-5_22

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free