Mapping Arabic WordNet synsets to Wikipedia articles using monolingual and bilingual features

Research output: Contribution to journalArticle

5 Citations (Scopus)

Abstract

The alignment of WordNet and Wikipedia has received wide attention from researchers of computational linguistics, who are building a new lexical knowledge source or enriching the semantic information of WordNet entities. The main challenge of this alignment is how to handle the synonymy and ambiguity issues in the contents of two units from different sources. Therefore, this paper introduces mapping method that links an Arabic WordNet synset to its corresponding article in Wikipedia. This method uses monolingual and bilingual features to overcome the lack of semantic information in Arabic WordNet. For evaluating this method, an Arabic mapping data set, which contains 1,291 synset–article pairs, is compiled. The experimental analysis shows that the proposed method achieves promising results and outperforms the state-of-the-art methods that depend only on monolingual features. The mapped method has also been used to increase the coverage of Arabic WordNet by inserting new synsets from Wikipedia.

Original languageEnglish
JournalNatural Language Engineering
DOIs
Publication statusAccepted/In press - 21 Oct 2015

Fingerprint

Wikipedia
Semantics
Computational linguistics
semantics
computational linguistics
WordNet
coverage
lack
knowledge

ASJC Scopus subject areas

  • Software
  • Artificial Intelligence
  • Language and Linguistics
  • Linguistics and Language

Cite this

@article{bfd76391d9c949d6be58a157521a3b9d,
title = "Mapping Arabic WordNet synsets to Wikipedia articles using monolingual and bilingual features",
abstract = "The alignment of WordNet and Wikipedia has received wide attention from researchers of computational linguistics, who are building a new lexical knowledge source or enriching the semantic information of WordNet entities. The main challenge of this alignment is how to handle the synonymy and ambiguity issues in the contents of two units from different sources. Therefore, this paper introduces mapping method that links an Arabic WordNet synset to its corresponding article in Wikipedia. This method uses monolingual and bilingual features to overcome the lack of semantic information in Arabic WordNet. For evaluating this method, an Arabic mapping data set, which contains 1,291 synset–article pairs, is compiled. The experimental analysis shows that the proposed method achieves promising results and outperforms the state-of-the-art methods that depend only on monolingual features. The mapped method has also been used to increase the coverage of Arabic WordNet by inserting new synsets from Wikipedia.",
author = "ABDULGABBAR SAIF and {Ab Aziz}, {Mohd Juzaiddin} and Nazlia Omar",
year = "2015",
month = "10",
day = "21",
doi = "10.1017/S1351324915000376",
language = "English",
journal = "Natural Language Engineering",
issn = "1351-3249",
publisher = "Cambridge University Press",

}

TY - JOUR

T1 - Mapping Arabic WordNet synsets to Wikipedia articles using monolingual and bilingual features

AU - SAIF, ABDULGABBAR

AU - Ab Aziz, Mohd Juzaiddin

AU - Omar, Nazlia

PY - 2015/10/21

Y1 - 2015/10/21

N2 - The alignment of WordNet and Wikipedia has received wide attention from researchers of computational linguistics, who are building a new lexical knowledge source or enriching the semantic information of WordNet entities. The main challenge of this alignment is how to handle the synonymy and ambiguity issues in the contents of two units from different sources. Therefore, this paper introduces mapping method that links an Arabic WordNet synset to its corresponding article in Wikipedia. This method uses monolingual and bilingual features to overcome the lack of semantic information in Arabic WordNet. For evaluating this method, an Arabic mapping data set, which contains 1,291 synset–article pairs, is compiled. The experimental analysis shows that the proposed method achieves promising results and outperforms the state-of-the-art methods that depend only on monolingual features. The mapped method has also been used to increase the coverage of Arabic WordNet by inserting new synsets from Wikipedia.

AB - The alignment of WordNet and Wikipedia has received wide attention from researchers of computational linguistics, who are building a new lexical knowledge source or enriching the semantic information of WordNet entities. The main challenge of this alignment is how to handle the synonymy and ambiguity issues in the contents of two units from different sources. Therefore, this paper introduces mapping method that links an Arabic WordNet synset to its corresponding article in Wikipedia. This method uses monolingual and bilingual features to overcome the lack of semantic information in Arabic WordNet. For evaluating this method, an Arabic mapping data set, which contains 1,291 synset–article pairs, is compiled. The experimental analysis shows that the proposed method achieves promising results and outperforms the state-of-the-art methods that depend only on monolingual features. The mapped method has also been used to increase the coverage of Arabic WordNet by inserting new synsets from Wikipedia.

UR - http://www.scopus.com/inward/record.url?scp=84944909488&partnerID=8YFLogxK

UR - http://www.scopus.com/inward/citedby.url?scp=84944909488&partnerID=8YFLogxK

U2 - 10.1017/S1351324915000376

DO - 10.1017/S1351324915000376

M3 - Article

JO - Natural Language Engineering

JF - Natural Language Engineering

SN - 1351-3249

ER -