Loading...
Search for: word-representations
0.011 seconds

    Persian word embedding evaluation benchmarks

    , Article 26th Iranian Conference on Electrical Engineering, ICEE 2018, 8 May 2018 through 10 May 2018 ; 2018 , Pages 1583-1588 ; 9781538649169 (ISBN) Zahedi, M. S ; Bokaei, M. H ; Shoeleh, F ; Yadollahi, M. M ; Doostmohammadi, E ; Farhoodi, M ; Sharif University of Technology
    Institute of Electrical and Electronics Engineers Inc  2018
    Abstract
    Recently, there has been renewed interest in semantic word representation also called word embedding, in a wide variety of natural language processing tasks requiring sophisticated semantic and syntactic information. The quality of word embedding methods is usually evaluated based on English language benchmarks. Nevertheless, only a few studies analyze word embedding for low resource languages such as Persian. In this paper, we perform such an extensive word embedding evaluation in Persian language based on a set of lexical semantics tasks named analogy, concept categorization, and word semantic relatedness. For these evaluation tasks, we provide three benchmark data sets to show the... 

    Persian language modeling using recurrent neural networks

    , Article 9th International Symposium on Telecommunication, IST 2018, 17 December 2018 through 19 December 2018 ; 2019 , Pages 207-210 ; 9781538682746 (ISBN) Hosseini Saravani, H ; Bahrani, M ; Veisi, H ; Besharati, S ; Sharif University of Technology
    Institute of Electrical and Electronics Engineers Inc  2019
    Abstract
    In this paper, recurrent neural networks are applied to language modeling of Persian, using word embedding as word representation. To this aim, unidirectional and bidirectional Long Short-Term Memory (LSTM) networks are used, and the perplexity of Persian language models on a 100-million-word data set is evaluated. The effect of various parameters, including number of hidden layers and size of LSTM units, on the performance of the networks in reducing the perplexity of the models are investigated. Among different LSTM language models, the best perplexity, which is equal to 59.05, is achieved from a 2-layer bidirectional LSTM model. Comparing this value with the perplexity of the classical...