Survey of BERT (Bidirectional Encoder Representation Transformer) types

19Citations
Citations of this article
70Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

There are many algorithms used in Natural Language Processing( NLP) to achieve good results, such as Machine Learning (ML), Deep Learning(DL) and many other algorithms. In Natural Language Processing,the first challenges is to convert text to numbers for using by any algorithm that a researcher choose. So how can convert text to numbers? This is happen by using Word Embedding algorithms such as skip gram,bags of words,BERT and etc. Representing words as numerical vectors by relying on the contents has become one of the effective methods for analyzing texts in machine learning, so that each word is represented by a vector to determine its meaning or to know how close or distant this word from the rest of the other word. BERT(Bidirectional Encoder Representation Transformer) is one of the embedding methods. It is designed to pre-trained form left and right in all layer deep training. It is a deep language model that is used for various tasks in natural language processing. In this paper we will review the different versions and types of BERT.

Cite

CITATION STYLE

APA

Mohammed, A. H., & Ali, A. H. (2021). Survey of BERT (Bidirectional Encoder Representation Transformer) types. In Journal of Physics: Conference Series (Vol. 1963). IOP Publishing Ltd. https://doi.org/10.1088/1742-6596/1963/1/012173

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free