Professional Documents
Culture Documents
• Both NER and Relation Extraction are foundational for many Natural
Language Processing (NLP) tasks.
• For Example :
❖ In the above Figure, we can see the NER process has detected that
Marie Curie, Pierre Curie, and Henri Becquerel are entities of type
PERSON, the Royal Swedish Academy of Sciences is an entity of type
ORGANISATION, and the Nobel Prize in Physics is an object (but
misclassified as a WORK_OF_ART).
• The NER process has many applications in NLP and search. For example,
❖ Identifying entities in search applications means that the user can now
search for “things, not strings,” and thereby the user can get more
relevant search results.
❖ NER can be applied to large collection of documents.
❖ NER can help reduce the unstructured text to a list of entities and their
counts, thereby allowing clustering or summarization to help humans
understand their contents.
❖ The output of NER can also be input into other processes, for example, a
coreference resolver that disambiguates the span “Marie” in “Marie and
Pierre Curie” to entity “Marie Curie.”
• Pretrained models in Named Entity Recognition (NER) are models that have
been trained on large amounts of text data to recognize and extract entities
such as person names, organizations, locations, and other entities from text.
• These models are usually trained using deep learning algorithms, such as
Convolutional Neural Networks (CNNs), Recurrent Neural Networks (RNNs),
or Transformers, on large datasets.
• Some popular pretrained NER models include the Stanford NER model, the
spaCy NER model, and the BERT-based NER models such as BioBERT,
SciBERT, and ClinicalBERT.
• Using a pretrained NER model can save time and effort in building an NER
system from scratch, and can also improve the accuracy of the system, as the
model has already learned to recognize and extract entities from text but
performance of a pretrained NER model may vary depending on the specific
domain and text data being analyzed.
• Sequence labelling is a Natural Language Processing task and has been one of
the most discussed topics in Linguistics and Computational Linguistics history.
• Sequence labeling can also be used for tasks such as Named Entity Recognition
(NER), Part-of-Speech (POS) tagging, and chunking.
• Information Extraction (IE) using sequence labeling is a popular approach in
natural language processing that involves tagging or labeling individual words
or tokens in a text sequence with different labels or tags that represent specific
information.
• In the context of IE, sequence labeling can be used to identify and extract
specific types of information from text.
• It aims to classify each token (word) in a class space C. This classification
approach can be independent (each word is treated independently) or
dependent (each word is dependent on other words).
• For example, in Named Entity Recognition, each word in a sentence is labeled
as representing a person, location, organization, or other entity. In relation
extraction, each word or phrase in a sentence is labeled with a tag that
indicates its role in the relationship between two entities.
• One of the most widely used sequence labeling algorithms in NLP is the
Hidden Markov Model (HMM), which models the probability distribution of
sequences of labels given a sequence of observations.
• Another popular approach is Conditional Random Fields (CRFs), which model
the conditional probability of label sequences given input sequences.
• Deep learning models such as Recurrent Neural Networks (RNNs) and
Transformers can also be used to get promising results for sequence labeling
tasks in NLP, especially in Named Entity Recognition.
• These models learn to capture the contextual information in the text sequence
by modeling the dependencies between each word or token and its surrounding
context.
• The main advantage of Sequence labeling is, its algorithms provide a
structured representation of text data that can be further processed and
analyzed as per requirement.
• For example, Part-Of-Speech Tagging (POS Tagging) is a sequence labeling
task. It is the process of marking up a word in a text (corpus) as corresponding
to a particular part of speech (syntactic tag), based on both its definition and its
context. POS Tagging is a helper task for many tasks about NLP: Word Sense
Disambiguation, Dependency Parsing, etc.
ISSUES IN MACHINE TRANSLATON Source : Open URL
• Personal pronouns : It means second person pronouns may vary in familiar and formal
situations.
3. Issues in Morphological analysis :
• For ex ; Chinese and Japanese do not use punctuations, it means sentences are not
separated by anything .
4. Issues in Syntactic Analysis :
• For ex; “ The boy saw a girl with a telescope “, means the girl had a telescope vs the
boy used a telescope to see the girl
5. Analysis of Context :
• Analysis of bilingual text corpora (source and target languages) and monolingual
corpora (target language) generates statistical models that transform text from one
language to another with that statistical weights are used to decide the most likely
translation.
• There are many SMT variants, depending upon how translation is modeled. Some
approaches are in terms of a string-to-string mapping, some use trees-to-strings, and
some use tree-to-tree models. All share in common the central idea that translation is
automatic, with models estimated from parallel corpora (source-target pairs) and also
from monolingual corpora.
• SMT is based on the idea that the probability of a given translation is dependent on
the probability of the source sentence and the target sentence.
• SMT works by analysing a large corpus of parallel texts, which are pairs of sentences in
the source and target languages that have been professionally translated. From this
corpus, SMT builds a statistical model that captures the probability of different
translations for a given sentence.
• The SMT system is then trained on this data to estimate the most likely translation of
an input sentence.
• The system does this by breaking down the input sentence into smaller units, such as
words or phrases, and then using the statistical model to estimate the probability of
different translations for each unit. The system then combines these translations to
produce a full translation of the input sentence.
5. Hybrid models:
• These models combine the strengths of multiple SMT models to improve
translation accuracy.
• For example, a hybrid model may use a phrase-based model for the initial
translation, followed by a syntax-based model to improve the grammatical
correctness of the output.
• Word alignment is a technique used in tasks such as machine translation and bilingual
text analysis.
• Bitext word alignment or simply word alignment is an important supporting task for
most methods of statistical machine translation.
• For example ; In the German translation of the line from “Smells Like Teen
Spirit” by Nirvana, the word “mich” doesn’t have a correspondent word in the
original version.
• Statistical machine translation models tunes their parameters based on the
word alignment.