Pay-Per-Request Deployment of Neural Network Models Using Serverless Architectures

20Citations
Citations of this article
80Readers
Mendeley users who have this article in their library.

Abstract

We demonstrate the serverless deployment of neural networks for model inferencing in NLP applications using Amazon’s Lambda service for feedforward evaluation and DynamoDB for storing word embeddings. Our architecture realizes a pay-per-request pricing model, requiring zero ongoing costs for maintaining server instances. All virtual machine management is handled behind the scenes by the cloud provider without any direct developer intervention. We describe a number of techniques that allow efficient use of serverless resources, and evaluations confirm that our design is both scalable and inexpensive.

Cite

CITATION STYLE

APA

Tu, Z., Li, M., & Lin, J. (2018). Pay-Per-Request Deployment of Neural Network Models Using Serverless Architectures. In NAACL HLT 2018 - 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Proceedings of the Demonstrations Session (pp. 6–10). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/n18-5002

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free