RMIT University
Browse

A Comprehensive Survey on Word Representation Models: From Classical to State-of-the-Art Word Representation Language Models

journal contribution
posted on 2024-11-02, 19:02 authored by Usman Naseem, Imran Razzak, Shah Khalid Khan, Mukesh Prasad
Word representation has always been an important research area in the history of natural language processing (NLP). Understanding such complex text data is imperative, given that it is rich in information and can be used widely across various applications. In this survey, we explore different word representation models and its power of expression, from the classical to modern-day state-of-the-art word representation language models (LMS). We describe a variety of text representation methods, and model designs have blossomed in the context of NLP, including SOTA LMs. These models can transform large volumes of text into effective vector representations capturing the same semantic information. Further, such representations can be utilized by various machine learning (ML) algorithms for a variety of NLP-related tasks. In the end, this survey briefly discusses the commonly used ML- and DL-based classifiers, evaluation metrics, and the applications of these word embeddings in different NLP tasks.

History

Related Materials

  1. 1.
    DOI - Is published in 10.1145/3434237
  2. 2.
    ISSN - Is published in 23754699

Journal

ACM Transactions on Asian and Low-Resource Language Information Processing

Volume

20

Number

74

Issue

5

Start page

1

End page

35

Total pages

35

Publisher

Association for Computing Machinery

Place published

United States

Language

English

Copyright

© 2021 Association for Computing Machinery.

Former Identifier

2006113040

Esploro creation date

2022-12-08

Usage metrics

    Scholarly Works

    Exports

    RefWorks
    BibTeX
    Ref. manager
    Endnote
    DataCite
    NLM
    DC