Computationally Efficient Neural Rendering for Generator Adversarial Networks Using a Multi-GPU Cluster in a Cloud Environment

7Citations
Citations of this article
19Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Due to its fantastic performance in the quality of the images created, Generator Adversarial Networks have recently become a viable option for image reconstruction. The main problem with employing GAN is how expensive the computations are. Researchers have developed techniques for distributing GANs across multiple nodes. However, these techniques typically do not scale because they frequently separate the components (Discriminator and Generator), leading to high communication overhead or encountering distribution-related problems unique to GAN training. In this study, the training procedure for the GAN is parallelized and carried out over many Graphical Processing Units (GPUs). TensorFlow's built-in logic and a custom loop were tweaked for more control over the resources allotted to each GPU worker. In this study, GPU image processing improvements and multi-GPU learning are used. The GAN model is accelerated using Distributed TensorFlow with synchronous data-parallel training on a single system and several GPUs. Acceleration was accomplished using the Genesis Cloud Platform and the NVIDIA GeForceTM GTX 108 GPU accelerator. The speed-up of 1.322 for two GPUs, 1.688 for three GPUs, and 1.7792 for four GPUs using multi-GPU acceleration. The parameter server model's data initialization and image production bottlenecks are removed, but the results' speed-up is not linear. Increasing the number of GPUs and removing the connectivity constraint will accelerate things even more. The bottlenecks are detected using new network lines and resources, and solutions are suggested. Recomputation and quantization are the two techniques to reduce the amount of GPU acceleration in memory. Deployment and versioning are essential for successfully operating multi-node GAN models in MLflow. Properly deploying and versioning these models can improve scalability, reproducibility, and collaboration across teams working on the same model. MLflow provides built-in tools for versioning and tracking model performance, making it easier to manage multiple versions of the model and reproduce it in different environments.

References Powered by Scopus

Detection of SARS-CoV-2 in Different Types of Clinical Specimens

4004Citations
N/AReaders
Get full text

A Saddle Point Algorithm for Networked Online Convex Optimization

130Citations
N/AReaders
Get full text

SCH-GAN: Semi-Supervised Cross-Modal Hashing by Generative Adversarial Network

111Citations
N/AReaders
Get full text

Cited by Powered by Scopus

Data augmentation and generative machine learning on the cloud platform

2Citations
N/AReaders
Get full text

A Systematic Review of Real-Time Deep Learning Methods for Image-Based Cancer Diagnostics

2Citations
N/AReaders
Get full text

Optimizing Target Recognition in Synthetic Aperture Radar Imagery: A Hyperparameter- Tuned Approach With Iterative Transfer Learning and Branched-Convolutional Neural Network

2Citations
N/AReaders
Get full text

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Cite

CITATION STYLE

APA

Ravikumar, A., & Sriraman, H. (2023). Computationally Efficient Neural Rendering for Generator Adversarial Networks Using a Multi-GPU Cluster in a Cloud Environment. IEEE Access, 11, 45559–45571. https://doi.org/10.1109/ACCESS.2023.3274201

Readers over time

‘23‘24‘25036912

Readers' Seniority

Tooltip

Lecturer / Post doc 1

50%

PhD / Post grad / Masters / Doc 1

50%

Readers' Discipline

Tooltip

Computer Science 2

100%

Article Metrics

Tooltip
Mentions
News Mentions: 2

Save time finding and organizing research with Mendeley

Sign up for free
0