Professional Documents
Culture Documents
N
q(zi) Пw=1VПK=1Tq(βwk) is the fully factorized variational
q(Z,B)=q(Z),q(B)=Пi=1
aapproximation to the posterior p(Z,β)is equal to minimization of KL
divergence between q(Z,β) and true posterior. Within this approximation, the
variational lower bound (L(q(Z),q(β),ө) takes the following form,
(L(q(Z),q(β),ө)=Eq[Σw=1VΣk=1Tlogp(βwk|α)-logq(βwk)+Σi=1N(logp(zi|xi,β)-
logq(zi)+Σj=1Clogp(yij|zi,xi,ө))]
Setting derivatives of L(q(Z),q(β),ө) w.r.to q(z) and q(β) to zero yields standard
update equations.
Fact extraction, coreference resolution etc are some of NLP tasks which highly
depends on existing word taxonomies or ontologies. The word taxonomy is
constructed by extracting taxonomical relations from a dictionary or
encyclopaedia. It consists of several relations. The quality of extracted
taxonomy is highly dependent on WSD results. Many of the WSD approaches is
modeled with ML tasks. The proposed work uses Lesk algorithm which is used
for feature representation highly rely on association, Word2Vec provides neural
network features and Adagram is used as word sense representation models. We
apply several WSD algorithms based on dictionary. Many work majorly
concentrate on the impact of different approaches to mine WSD features in
order to improve word frequency. WSD detects the precise sense from the set of
all possible senses based on contexts. The noun is a ambiguous word have
multiple senses. Hypernym represents “Is a” relationship which comes under
noun word. The challenge associated with WSD is the generation of word
taxonomy.
In word taxonomy, a directed graph consists of nodes represents word senses
and edges represents hyponym-hypernym relation. For e.g apple is a fruit. Here
apple is a specific word. Fruit is a generic word. The semantic resources are
lexical database, thesauri, ontologies etc. Part of NLP tasks are constructing a
database of semantic similarity or a tool for term generalization. There are
several kinds of approaches to create or update a taxonomy. Word taxonomy
can be
i)generated manually by lexicographers.
ii) converted from existing structured resource
iii)extracted from a corpus, or derived from a corpus
iv)derived from corpus trained vector semantic model.
Corpus extraction efficiency and methods vary greatly based on corpus with
remarkable works performed on corpora of dictionary glosses, formal text
corpora and large general corpora. Each of these approaches is a trade-off
between required labour and quality of resulting taxonomy.
As a consequence for taxonomy extraction ;a monolingual dictionary consists of
sentences that contain hypernym. Hypernym occupies same syntactic position.It
is able to generate high quality taxonomies by mining hypernym relations from
corpora. This work is bound to extracting taxonomy from Russian language
from monolingual dictionary. WSD methods from general corpora are fitted for
this kind of WSD tasks. This paper focuses on application of WSD methods in
hypernym disambiguation in monolingual dictionary. It also describes the
parameters of WSD methods that are helpful to solve WSD problem.
Organization of paper is aligned as
1. Brief overview of existing approaches to the problem of WSD.
2. A description of data sources, data preparation and annotation.
3. A description of WSD pipeline that compares different feature extraction
and machine learning configurations.
4. Description and analysis of WSD parameters and their performance
5. Discussion of results and finally concluding the activity.
Background
All the approaches related to WSD is based on context that defines word sense
though the context varies.
Lesk algorithm measures the similarity metric between two contexts.Even it
perform well it suffers from data sparcity.The simplest solution is to overcome
the limitation to use semantic relatedness databases.The usage of WordNet
synset is to add more overlapping context words.Siddorov et al increases the
matches between two context using extended synonym dictionary and a
dedicated derivation morphology system.It has high WSD precision on a
Spanish corpus.Many more attempts are uses ML in a WSD task
e.g.LDA,Maximum entropy classifier ,genetic algorithms and others.Approach
related to WSD based on neural network with auto encoder or similar topology
.However early approaches are high computation demand and slow and noisy
learning algorithms.Mikolov et al trained autoencoder ,Word2Vec shows
similarity between arithmetic operations on autoencoder derived word
embedding and semantic relations.It insists Skip gram model is better than
CBOW model.Word embedding model doesnot provide a singly way to word
context into feature vector.It also composed Word embedding features from the
corpus for WSD.They tested several representation of sense;a concatenation or
different weighted averages of word context vectors.Many works were carried
out to construct word embedding model whose senses are assigned to vectors.It
sg model ; a semantically disambiguated model.It uses resulting set of vectors
as a semantic relatedness databased in a WSD tasks.Chen et al uses iteratively
performing WSD on a corpus using Skip-gram model and training the model on
a resultant corpus to improve SG model train the model on a resultant corpus to
improve WSD performance over naïve SG model.It is very demanding in both
time and space requirement.The practical implementation of WSD is the direct
induction of WS embedding.Adagram;a non-parametic extension of SG model
that conducts Bayesian induction of word senses and optimization of word
senses embedding representations and word sense probabilities in a given
context.RNN in NLP approaches WSD based on LSTM model uses coarse
model of how human reads sentences sequentially.
[J] Word Sense Disambiguation is widely used in NLP tasks such as Natural
Language Processing,Machine Learning .The experimental analysis of the
proposed work includes context mining ,feature analysis and text
classification.Adjective plays a vital part in text classification using machine
learning algorithms.The application of the proposed work includes document
indexing based on controlled vocabulary,adjective,word sense
disambiguation,constructing hierarchical categorization of web pages,spam
detection,topic labelling,web search,document summairization etc..Feature
extraction is characterized by using cuckoo search algorithm and text
classification is performed by using linear support vector machine.Text
document mining uses both machine learning algorithm and deep neural
network.WSD detect exact sense to the intended word that distinguish from all
other possible senses.Text document mining removes irrelevant ,redundant and
noisy features.FE and FS are two techniques used for FR techniques.FE
technique is characterized from new and low-dimensional element space.E.g.
PCA,LDA etc.Feature Analysis techniques assign weights in the range of (0,1)
to each feature.The WSD identifies the sense of words.Knowledge based
frameworks highlights the importance of words of last SenseEval event