A study of performance scalability by parallelizing loop iterations on multi-core SMPs

1Citations
Citations of this article
1Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Today, the challenge is to exploit the parallelism available in the way of multi-core architectures by the software. This could be done by re-writing the application, by exploiting the hardware capabilities or expect the compiler/software runtime tools to do the job for us. With the advent of multi-core architectures ([1] [2]), this problem is becoming more and more relevant. Even today, there are not many run-time tools to analyze the behavioral pattern of such performance critical applications, and to re-compile them. So, techniques like OpenMP for shared memory programs are still useful in exploiting parallelism in the machine. This work tries to study if the loop parallelization (both with and without applying transformations) can be a good case for running scientific programs efficiently on such multi-core architectures. We have found the results to be encouraging and we strongly feel that this could lead to some good results if implemented fully in a production compiler for multi-core architectures. © Springer-Verlag Berlin Heidelberg 2010.

Cite

CITATION STYLE

APA

Raghavendra, P., Behki, A. K., Hariprasad, K., Mohan, M., Jain, P., Bhat, S. S., … Prabhu, V. (2010). A study of performance scalability by parallelizing loop iterations on multi-core SMPs. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 6081 LNCS, pp. 476–486). https://doi.org/10.1007/978-3-642-13119-6_41

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free