abide by the grammatical syntax of ISL and then
mapped to the dictionary of the video animations.
Words not present are broken into letters and shown
one by one. These can be for things such as a name or
a place.
The method includes displaying sign language
using an avatar after translating SiGML to motion
data.
A major challenge in the system is the conversion
of one language to another with a completely different
set of grammatical syntax in place.
Also, the feature of text summarization has been
added to deliver large volume of data in smaller
amount of time.
1.2 Text Summarization
There is enormous volume of textual content that is
generated on the Internet and in the numerous
archives of headlines, academic papers, government
documentation, etc., automatic text summarization
(ATS) has become more crucial. With the enormous
amount of textual content, manual text summarizing
takes a lot of time, effort, money, and sometimes
impracticable (El-Kassas et al., 2021). A variety of
tasks can be done using ATS like generation of
summary for a scientific paper, news articles, creating
summary of audio podcast etc.
A strategy to extracting highlights based on a
recognized contextualized embedding architecture
(Moreno et al., 2022)., especially the transformers, is
known as a Transformer-based Highlights Extractor
(THExt, in short). BART is a sequence-to-sequence
model trained as a denoising autoencoder, It is
applicable to many types of tasks like sequence
classification (categorizing input text sentences or
tokens), summarizing text, machine translation like
translation between multiple language, question
answering. Its pretraining has mainly two phases.
Assign corrupted text with an arbitrary noise and
sequence-to-sequence model is learned to rebuild the
actual text. It is evaluated with a different noise
approach as shown in Figure 1, like randomly
shuffling the order of the original text and using a
novel in-filling scheme (in this scheme length of span
of text are replaced with mask token). It is an
unsupervised language model which can be fine-
tuned to a specific application like medical chatbots,
generating summary of meeting, natural language to
programming language, language translation etc. As
it is already pretrained with very large amount of data,
a small data set can be used to fine-tune it.
Figure 1: Transformations for noising the input for BART
fine-tune.
By using the BART transformer model, we can
automate the text summarization task. Text
summarization can be done in 2 ways.
Extractive summarization: It provides the
important text present in the given input.
Abstractive summarization: It provides the
actual summary of the given input. So, it is
more challenging as it has to read complete text
and understand the meaning of the text and
provide us with the summary.
1.3 Text Translation
The ability to translate moderate languages has
significantly improved due to training a universal
translation system between different languages (Firat
et al., 2016).
Recent research (Arivazhagan et al.,
2019; Fan et al., 2020) has also shown that
multilingual translation models in a single model
have a great potential for performance improvement.
Using the pretrain and fine-tune approach common to
NLP, recent pretrained multilingual sequence-to-
sequence (seq2seq) models have made it simple to
build neural machine translation (MT) systems (Liu
et al., 2020). Pretrained models are excellent
candidates for MT domain adaption tasks, where
domain-specific bitext is typically less accessible
than general bitext, because fine-tuning these models
typically requires less data than is required for from-
scratch translation models.
For the translation of text from any language to
English, we are using mBART (Yuqing et al., 2020).
It is a multilingual neural machine translation model.
mBART supports up to 50 languages. Initially, the
mBART model trains in 25 different languages and is
fine-tuned for different tasks. For translation, it was
fine-tuned on bitext (bilingual finetuning). Later in
mBART50, the multilingual nature of the pretraining
is used for fine-tuning the model.