have developed specific methods for setting up and
evaluating accent analysis. Accents vary in terms of
sound quality, phoneme articulation and prosody.
Because it is challenging to extract these precise
features, they concluded that existing work uses
alternative features, such as spectral features that
capture the frequency of speech, including five
significant features: Mel Falling Spectral Coefficient
(MFCC), spectrogram, chromaticity map, spectral
centroid and spectral roll-off (Shergill et al. 2020).
They have researched and proposed a new way of
capturing accents to improve the recognition
efficiency of the system. These features can improve
the accuracy of accent categorisation for accented
language systems. Other researchers have categorised
user input capture in the form of speech data. They
are modelling human speech accents and gender
recognition as classification tasks. They used a deep
convolutional neural network and experimentally
developed an architecture that maximises
classification accuracy for the above functions.
Gender categorisation was more straightforward to
predict with high accuracy than accent. The
categorisation of accents was complex because the
overlap of regional accents prevented them from
being categorised with high accuracy (Najafian et al.
2016). Deng et al. have carefully delineated the field
of automatic speech recognition within a framework
such as improved accent recognition and accented
speech recognition within a self-supervised learning
framework (Deng et al. 2021). Some other studies
have categorised accents, particularly the origin of
English accents, to classify them for better
identification (Ai et al 2008).
This paper presents a comprehensive review of
the application of artificial intelligence models in
accented speech recognition systems. Section 2
outlines the workflow of various relevant
methodologies. A discussion on the findings and
implications of these applications is provided in
Section 3. The paper concludes in Section 4 with a
summary of the key insights and contributions of the
research.
2 METHOD
2.1 Framework of Developing Machine
Learning Model
A prevailing notion within the philosophy of science
posits that models which are simplified and idealized
are more comprehensible than their complex or
abstract counterparts (Bokulich 2008). The landscape
of artificial intelligence (AI) architecture is
continuously evolving, leading to the development of
numerous machine learning models, with simpler
models often being more straightforward to grasp.
These models are particularly adept at addressing
"what-if" scenarios or exploring causal relationships,
thereby effectively identifying and underscoring
critical distinctions. Deep Neural Networks (DNNs)
serve as a prime example of this. They fundamentally
operate by leveraging vast datasets to perform
classifications, predictions, and inferences.
Specifically, DNNs process extensive data inputs to
produce representations that facilitate generalized
predictions, enabling the anticipation of future events
(Sullivan 2022). The most essential steps in machine
modelling are model building, data collection and
model management. Building machine learning
models is an iterative process. Data collection is
becoming one of the critical bottlenecks. It is well
known that running machine learning end-to-end,
including collection, cleaning, analysis, visualisation
and feature engineering. But one problem that arises
with the emergence of new machine learning models
is the lack of data, especially the lack of training data
such as Deep learning, which may require a lot of
training data to perform well. The simple method of
manual labelling can be used when there is a lack of
training data, but it is expensive and requires domain
expertise (Roh et al. 2021). There are three main
approaches to data collection. First, if the goal is to
share and search for new datasets, data collection
techniques can be used to discover, extend, or
generate datasets. Second, once a dataset is available,
various data labelling techniques can be used to
annotate individual examples. Finally, rather than
labelling new datasets, existing data can be improved
or trained on well-trained models. These three
approaches are not necessarily distinct and can be
used simultaneously (Roh et al. 2021).
Management of the machine learning model is
equally important. It is impossible to manage models
that have moved on over time sustainably. At the
same time, the management of the model is equally
important. For example, Manasi Vartak et al.
contributed to model management by developing a
new software management model database for
management. To automate the tracking of machine
learning models, the back-end introduces a standard
abstraction layer to represent models and pipelines.
The ModelDB front-end allows for visual exploration
and analysis of models through a web-based platform.
The management of machine learning models is
achieved by visual exploration and analysis of models
through a web-based interface (Vartak et al. 2016).